2025-04-02 02:16:59,486 [ 51455 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2025-04-02 02:16:59,487 [ 51455 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:97, check_args_and_update_paths) 2025-04-02 02:16:59,487 [ 51455 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:108, check_args_and_update_paths) 2025-04-02 02:16:59,487 [ 51455 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:110, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_8ua42k --privileged --dns-search='.' --memory=30709030912 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=8b2301119731 -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e CLICKHOUSE_USE_OLD_ANALYZER=1 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_aggregation_memory_efficient/test.py::test_remote test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile test_analyzer_compatibility/test.py::test_two_new_versions test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables test_attach_table_normalizer/test.py::test_attach_substr test_attach_table_normalizer/test.py::test_attach_substr_restart test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 test_config_corresponding_root/test.py::test_work test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed test_config_substitutions/test.py::test_allow_databases test_config_substitutions/test.py::test_config test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config 'test_database_backup/test.py::test_database_backup_database[Disk('\"'\"'backup_disk_local'\"'\"', '\"'\"'test_database_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('\"'\"'backup_disk_object_storage_local_plain'\"'\"', '\"'\"'test_database_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('\"'\"'backup_disk_s3_plain'\"'\"', '\"'\"'test_database_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_database[File('\"'\"'test_database_backup_file'\"'\"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('\"'\"'backup_disk_local'\"'\"', '\"'\"'test_table_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('\"'\"'backup_disk_object_storage_local_plain'\"'\"', '\"'\"'test_table_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('\"'\"'backup_disk_s3_plain'\"'\"', '\"'\"'test_table_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_table[File('\"'\"'test_table_backup_file'\"'\"')]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Date_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[String_]' -vvv -ss" altinityinfra/integration-tests-runner:2165613c5fcd '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache Test order randomisation NOT enabled. Enable with --random-order or --random-order-bucket= rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.3.1, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0, random-order-1.1.1 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [100 items] scheduling tests via LoadFileScheduling test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] Command:[docker ps | wc -l] Command:[docker ps | wc -l] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] test_async_load_databases/test.py::test_async_load_system_database Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] test_backup_restore/test.py::test_attach_partition test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings Command:[docker ps | wc -l] Command:[docker ps | wc -l] test_cluster_all_replicas/test.py::test_cluster test_config_substitutions/test.py::test_allow_databases test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Pruning Docker networks Command:[docker network prune --force] Stdout:1 Stdout:1 No running containers Pruning Docker networks No running containers Pruning Docker networks Command:[docker network prune --force] Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Stdout:1 Pruning Docker networks No running containers Command:[docker network prune --force] Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Stdout:net.ipv4.ip_local_port_range = 55000 65535 Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Stdout:net.ipv4.ip_local_port_range = 55000 65535 Source RedisHash incompatible with layout direct Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 133f11833cd2 ENV SHLVL 0 Source RedisSimple incompatible with layout complex_key_hashed ENV HOME /root ENV OLDPWD / Source RedisSimple incompatible with layout complex_key_cache ENV DOCKER_HELPER_TAG 5dc43a6382f0 Source RedisSimple incompatible with layout complex_key_direct ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache ENV UBSAN_OPTIONS print_stacktrace=1 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_aggregation_memory_efficient/test.py::test_remote test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile test_analyzer_compatibility/test.py::test_two_new_versions test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables test_attach_table_normalizer/test.py::test_attach_substr test_attach_table_normalizer/test.py::test_attach_substr_restart test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 test_config_corresponding_root/test.py::test_work test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed test_config_substitutions/test.py::test_allow_databases test_config_substitutions/test.py::test_config test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_local'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[File('"'"'test_database_backup_file'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_local'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[File('"'"'test_table_backup_file'"'"')]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Date_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[String_]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 Source RedisHash incompatible with layout direct ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV CLICKHOUSE_USE_OLD_ANALYZER 1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 Running tests in /ClickHouse/tests/integration/test_backup_restore_on_cluster/test.py Running tests in /ClickHouse/tests/integration/test_cluster_all_replicas/test.py ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 Cluster start called. is_up=False ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 Cluster start called. is_up=False ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Source RedisSimple incompatible with layout complex_key_hashed ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config Source RedisSimple incompatible with layout complex_key_cache ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat ENV DOCKER_BASE_TAG 8b2301119731 Source RedisHash incompatible with layout hashed ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 Source RedisHash incompatible with layout cache ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache ENV WORKER_FREE_PORTS 30250 30251 30252 30253 30254 30255 30256 30257 30258 30259 30260 30261 30262 30263 30264 30265 30266 30267 30268 30269 30270 30271 30272 30273 30274 30275 30276 30277 30278 30279 30280 30281 30282 30283 30284 30285 30286 30287 30288 30289 30290 30291 30292 30293 30294 30295 30296 30297 30298 30299 Source RedisSimple incompatible with layout complex_key_direct ENV PYTEST_XDIST_TESTRUNUID 403ebba7b09548d7ae9e72b89a600d03 Stdout:net.ipv4.ip_local_port_range = 55000 65535 Source RedisHash incompatible with layout flat ENV PYTEST_XDIST_WORKER gw5 Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] (setup) Running tests in /ClickHouse/tests/integration/test_backup_restore/test.py Source RedisHash incompatible with layout direct Cluster start called. is_up=False CLUSTER INIT base_config_dir:/clickhouse-config Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Stdout:net.ipv4.ip_local_port_range = 55000 65535 Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Running tests in /ClickHouse/tests/integration/test_async_load_databases/test.py Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Cluster start called. is_up=False Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Stdout:net.ipv4.ip_local_port_range = 55000 65535 Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Running tests in /ClickHouse/tests/integration/test_config_substitutions/test.py Cluster start called. is_up=False Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Stdout:net.ipv4.ip_local_port_range = 55000 65535 Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Running tests in /ClickHouse/tests/integration/test_database_backup/test.py Cluster start called. is_up=False Source RedisHash incompatible with layout direct clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Source RedisSimple incompatible with layout complex_key_hashed Cluster name:mysql project_name:roottestdictionariesalllayoutsseparatesourcesmysql-gw5. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/.env', '--project-name', 'roottestdictionariesalllayoutsseparatesourcesmysql-gw5', '--file', '/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Running tests in /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/test_mysql.py Source RedisHash incompatible with layout direct Cluster start called. is_up=False Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_UInt8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_UInt8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_UInt8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_UInt32_.xml Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Float64_.xml Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_UInt8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_UInt8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Float64_.xml Stdout:net.ipv4.ip_local_port_range = 55000 65535 Docker networks for project roottestclusterallreplicas-gw4 are NETWORK ID NAME DRIVER SCOPE Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_UInt8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_DateTime_.xml Running tests in /ClickHouse/tests/integration/test_allow_feature_tier/test.py Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_DateTime_.xml Cluster start called. is_up=False Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_UInt8_.xml clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name: project_name:roottestdictionariesredis-gw2. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/.env', '--project-name', 'roottestdictionariesredis-gw2', '--file', '/ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Running tests in /ClickHouse/tests/integration/test_dictionaries_redis/test.py Cluster start called. is_up=False Stdout:net.ipv4.ip_local_port_range = 55000 65535 Docker networks for project roottestbackuprestore-gw6 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestbackuprestoreoncluster-gw0 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestdictionariesalllayoutsseparatesourcesmysql-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestdatabasebackup-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestasyncloaddatabases-gw9 are NETWORK ID NAME DRIVER SCOPE ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 133f11833cd2 ENV SHLVL 0 ENV HOME /root ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV UBSAN_OPTIONS print_stacktrace=1 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_aggregation_memory_efficient/test.py::test_remote test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile test_analyzer_compatibility/test.py::test_two_new_versions test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables test_attach_table_normalizer/test.py::test_attach_substr test_attach_table_normalizer/test.py::test_attach_substr_restart test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 test_config_corresponding_root/test.py::test_work test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed test_config_substitutions/test.py::test_allow_databases test_config_substitutions/test.py::test_config test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_local'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[File('"'"'test_database_backup_file'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_local'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[File('"'"'test_table_backup_file'"'"')]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Date_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[String_]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV CLICKHOUSE_USE_OLD_ANALYZER 1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_BASE_TAG 8b2301119731 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 Docker networks for project roottestconfigsubstitutions-gw7 are NETWORK ID NAME DRIVER SCOPE ENV INTEGRATION_TESTS_RUN_ID 0 ENV WORKER_FREE_PORTS 30150 30151 30152 30153 30154 30155 30156 30157 30158 30159 30160 30161 30162 30163 30164 30165 30166 30167 30168 30169 30170 30171 30172 30173 30174 30175 30176 30177 30178 30179 30180 30181 30182 30183 30184 30185 30186 30187 30188 30189 30190 30191 30192 30193 30194 30195 30196 30197 30198 30199 ENV PYTEST_XDIST_TESTRUNUID 403ebba7b09548d7ae9e72b89a600d03 ENV PYTEST_XDIST_WORKER gw3 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] (setup) CLUSTER INIT base_config_dir:/clickhouse-config clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name:clickhouse_local project_name:roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3. Added instance name:local_node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/.env', '--project-name', 'roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3', '--file', '/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/local_node/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Running tests in /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py Cluster start called. is_up=False Docker containers for project roottestclusterallreplicas-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestbackuprestore-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker networks for project roottestallowfeaturetier-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestdictionariesredis-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackuprestoreoncluster-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestasyncloaddatabases-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestdictionariesalllayoutsseparatesourcesmysql-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker networks for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfigsubstitutions-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestdatabasebackup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestdictionariesredis-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackuprestoreoncluster-gw0 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestclusterallreplicas-gw4 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestbackuprestore-gw6 are DRIVER VOLUME NAME Cleanup called Docker containers for project roottestallowfeaturetier-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestdictionariesalllayoutsseparatesourcesmysql-gw5 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestasyncloaddatabases-gw9 are DRIVER VOLUME NAME Cleanup called Docker containers for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigsubstitutions-gw7 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestbackuprestoreoncluster-gw0 are NETWORK ID NAME DRIVER SCOPE Docker volumes for project roottestdictionariesredis-gw2 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestdatabasebackup-gw1 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestclusterallreplicas-gw4 are NETWORK ID NAME DRIVER SCOPE Docker volumes for project roottestallowfeaturetier-gw8 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestbackuprestore-gw6 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestdictionariesalllayoutsseparatesourcesmysql-gw5 are NETWORK ID NAME DRIVER SCOPE Docker volumes for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestdictionariesredis-gw2 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestdatabasebackup-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestasyncloaddatabases-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestconfigsubstitutions-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestallowfeaturetier-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackuprestoreoncluster-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestclusterallreplicas-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestbackuprestore-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestdatabasebackup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestdictionariesredis-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestallowfeaturetier-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestconfigsubstitutions-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestdictionariesalllayoutsseparatesourcesmysql-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestasyncloaddatabases-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackuprestoreoncluster-gw0 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestoreoncluster-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker networks for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 are NETWORK ID NAME DRIVER SCOPE Docker volumes for project roottestbackuprestore-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestore-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestclusterallreplicas-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestclusterallreplicas-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestdatabasebackup-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdatabasebackup-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestconfigsubstitutions-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestdictionariesredis-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdictionariesredis-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestconfigsubstitutions-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestallowfeaturetier-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestallowfeaturetier-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackuprestoreoncluster-gw0 Trying to prune unused networks... Docker volumes for project roottestasyncloaddatabases-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestasyncloaddatabases-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestdictionariesalllayoutsseparatesourcesmysql-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdictionariesalllayoutsseparatesourcesmysql-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker containers for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Unstopped containers: {} No running containers for project: roottestallowfeaturetier-gw8 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestbackuprestore-gw6 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestconfigsubstitutions-gw7 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestdictionariesredis-gw2 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestclusterallreplicas-gw4 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestdatabasebackup-gw1 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Unstopped containers: {} No running containers for project: roottestasyncloaddatabases-gw9 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestdictionariesalllayoutsseparatesourcesmysql-gw5 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Docker volumes for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Trying to prune unused images... Stdout:Total reclaimed space: 0B Command:[docker image prune -f] Trying to prune unused images... Images pruned Trying to prune unused volumes... Command:[docker image prune -f] Command:[docker volume ls | wc -l] Trying to prune unused images... Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Exitcode:1 Trying to prune unused volumes... Stderr:Error response from daemon: a prune operation is already running Command:[docker volume ls | wc -l] Exitcode:1 Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Trying to prune unused volumes... Command:[docker volume ls | wc -l] Unstopped containers: {} No running containers for project: roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 Trying to prune unused networks... Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Stdout:1 Volumes pruned: 1 Command:[docker volume ls | wc -l] Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster3.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_access_storage.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_user_defined_sql_objects.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/lesser_timeouts.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/database Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/backups Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Copy common configuration from helpers Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_async_load_databases/configs/config.xml'] to /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node1/configs/config.d Stdout:1 Stdout:1 Volumes pruned: 1 Setup directory for instance: instance Stdout:1 Volumes pruned: 1 Setup directory for instance: instance Volumes pruned: 1 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster3.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_access_storage.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_user_defined_sql_objects.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/lesser_timeouts.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/configs/config.d Create directory for configuration generated in this helper Create directory for common tests configuration Setup database dir /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node1/database Copy common configuration from helpers Generate and write macros file Generate and write macros file Setup logs dir /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node1/logs Copy custom test config files ['/ClickHouse/tests/integration/test_database_backup/configs/backups.xml'] to /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/configs/config.d Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Copy custom test config files ['/ClickHouse/tests/integration/test_allow_feature_tier/configs/allow_feature_tier.xml'] to /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/configs/config.d Generate and write macros file Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/database Setup directory for instance: node2 Copy custom test config files [] to /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/node/configs/config.d Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/backups Setup directory for instance: node3 Setup database dir /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/database Create directory for configuration generated in this helper Create directory for common tests configuration Setup logs dir /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Copy common configuration from helpers Create directory for configuration generated in this helper Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env Create directory for common tests configuration Copy common configuration from helpers Setup database dir /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/database Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Generate and write macros file No config file found Setup logs dir /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/logs Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Copy custom test config files ['/ClickHouse/tests/integration/test_async_load_databases/configs/async_load_system_database.xml'] to /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node2/configs/config.d Generate and write macros file No config file found Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster3.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_access_storage.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_user_defined_sql_objects.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/lesser_timeouts.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/configs/config.d Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Setup database dir /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node2/database Setup logs dir /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node2/logs Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/database Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/.env external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/backups Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Stdout:1 Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_cluster_all_replicas/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/node/database Setup logs dir /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Stdout:1 Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'REDIS_HOST': 'redis1', 'REDIS_EXTERNAL_PORT': '30100', 'REDIS_INTERNAL_PORT': '6379'} stored in /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/.env Volumes pruned: 1 Setup directory for instance: node Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Create directory for configuration generated in this helper Setup database dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/database Setup logs dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/logs Create directory for common tests configuration Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Copy common configuration from helpers Setup directory for instance: node2 Stdout:1 Volumes pruned: 1 Create directory for configuration generated in this helper Generate and write macros file Create directory for common tests configuration Setup directory for instance: node1 Copy common configuration from helpers Copy custom test config files [] to /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/node/database Create directory for configuration generated in this helper Setup logs dir /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/node/logs Create directory for common tests configuration Generate and write macros file Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Copy common configuration from helpers Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/.env Copy custom test config files ['/ClickHouse/tests/integration/test_cluster_all_replicas/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/database Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Setup logs dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/logs Generate and write macros file Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env Copy custom test config files [] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node1/configs/config.d No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node1/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 http://localhost:None "GET /version HTTP/1.1" 200 826 Create directory for configuration generated in this helper Command:[docker compose --env-file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env --project-name roottestallowfeaturetier-gw8 --file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/docker-compose.yml pull] Create directory for common tests configuration Copy common configuration from helpers Trying to prune unused images... http://localhost:None "GET /version HTTP/1.1" 200 826 Generate and write macros file Command:[docker image prune -f] Copy custom test config files [] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node2/configs/config.d Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --project-name roottestbackuprestoreoncluster-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/docker-compose.yml pull] Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node2/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node2/logs http://localhost:None "GET /version HTTP/1.1" 200 826 Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node3 Command:[docker compose --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --project-name roottestdatabasebackup-gw1 --file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml pull] Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/config_zk_include_test.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node3/configs/config.d http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/.env --project-name roottestdictionariesredis-gw2 --file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml pull] Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node3/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node3/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Setup directory for instance: node4 http://localhost:None "GET /version HTTP/1.1" 200 826 Create directory for configuration generated in this helper Create directory for common tests configuration Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/.env --project-name roottestasyncloaddatabases-gw9 --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node2/docker-compose.yml pull] Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/include_from_source.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node4/configs/config.d Stdout:1 Volumes pruned: 1 Setup directory for instance: node Create directory for configuration generated in this helper Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node4/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node4/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Create directory for common tests configuration Setup directory for instance: node5 Copy common configuration from helpers Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node5/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/configs/disable_ssl_verification.xml'] to /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node5/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node5/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node6 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Setup database dir /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/node/database Generate and write macros file Setup logs dir /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/node/logs Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/include_from_source.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node6/configs/config.d Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MYSQL8_HOST': 'mysql80', 'MYSQL8_PORT': '3306', 'MYSQL8_ROOT_HOST': '%', 'MYSQL8_LOGS': '/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/mysql8/logs', 'MYSQL8_LOGS_FS': 'bind', 'MYSQL8_DOCKER_USER': '0'} stored in /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/.env Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node6/database Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node6/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node7 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/000-server_overrides.xml', '/ClickHouse/tests/integration/test_config_substitutions/configs/010-server_with_env_subst.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node7/configs/config.d http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env --project-name roottestclusterallreplicas-gw4 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/docker-compose.yml pull] Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node7/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node7/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] http://localhost:None "GET /version HTTP/1.1" 200 826 Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MAX_QUERY_SIZE': '121212', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper3/coordination', 'INCLUDE_FROM_ENV': '/etc/clickhouse-server/config.d/include_from_source.xml', 'MAX_THREADS': '2'} stored in /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node7/.env Setup directory for instance: node8 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/.env --project-name roottestbackuprestore-gw6 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/node/docker-compose.yml pull] Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/include_from_source.yml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node8/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node8/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node8/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MAX_QUERY_SIZE': '55555', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper3/coordination', 'INCLUDE_FROM_ENV': '/etc/clickhouse-server/config.d/include_from_source.xml'} stored in /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/.env --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw5 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml pull] http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/.env --project-name roottestconfigsubstitutions-gw7 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node8/docker-compose.yml pull] Stdout:1 Volumes pruned: 1 Setup directory for instance: local_node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/configs/disable_ssl_verification.xml'] to /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/local_node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/local_node/database Setup logs dir /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/local_node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/.env --project-name roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/local_node/docker-compose.yml pull] Stderr: node2 Skipped - Image is already being pulled by node1 Stderr: node1 Pulling Stderr: node1 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env --project-name roottestclusterallreplicas-gw4 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env --project-name roottestclusterallreplicas-gw4 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate] Stderr: zoo2 Skipped - Image is already being pulled by zoo1 Stderr: zoo3 Skipped - Image is already being pulled by zoo1 Stderr: node2 Skipped - Image is already being pulled by zoo1 Stderr: node3 Skipped - Image is already being pulled by zoo1 Stderr: node1 Skipped - Image is already being pulled by zoo1 Stderr: zoo1 Pulling Stderr: zoo1 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/coordination', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/coordination', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/coordination'] Command:[docker compose --project-name roottestbackuprestoreoncluster-gw0 --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: node1 Skipped - Image is already being pulled by node2 Stderr: node2 Pulling Stderr: node2 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/.env --project-name roottestasyncloaddatabases-gw9 --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/.env --project-name roottestasyncloaddatabases-gw9 --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node2/docker-compose.yml up -d --no-recreate] Stderr: instance Pulling Stderr: instance Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env --project-name roottestallowfeaturetier-gw8 --file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env --project-name roottestallowfeaturetier-gw8 --file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/docker-compose.yml up -d --no-recreate] Stderr: local_node Pulling Stderr: local_node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/.env --project-name roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/local_node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/.env --project-name roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/local_node/docker-compose.yml up -d --no-recreate] Stderr: zoo3 Skipped - Image is already being pulled by node8 Stderr: node5 Skipped - Image is already being pulled by node8 Stderr: node3 Skipped - Image is already being pulled by node8 Stderr: node1 Skipped - Image is already being pulled by node8 Stderr: node6 Skipped - Image is already being pulled by node8 Stderr: node2 Skipped - Image is already being pulled by node8 Stderr: node7 Skipped - Image is already being pulled by node8 Stderr: node4 Skipped - Image is already being pulled by node8 Stderr: zoo1 Skipped - Image is already being pulled by node8 Stderr: zoo2 Skipped - Image is already being pulled by node8 Stderr: node8 Pulling Stderr: node8 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper1/log', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper1/config', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper1/coordination', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper2/log', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper2/config', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper2/coordination', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper3/log', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper3/config', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/keeper3/coordination'] Command:[docker compose --project-name roottestconfigsubstitutions-gw7 --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: proxy1 Pulling Stderr: resolver Pulling Stderr: instance Pulling Stderr: minio1 Pulling Stderr: minio1 Pulled Stderr: resolver Pulled Stderr: instance Pulled Stderr: proxy1 Pulled Trying to create Minio instance by command docker compose --project-name roottestdatabasebackup-gw1 --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Command:[docker compose --project-name roottestdatabasebackup-gw1 --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] Stderr:time="2025-04-02T02:17:17Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestbackuprestoreoncluster-gw0_default Creating Stderr: Network roottestbackuprestoreoncluster-gw0_default Created Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Started Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Started Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Started Stderr:time="2025-04-02T02:17:18Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:17:18Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.1.4, port:2181, use_ssl:False Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stderr: Network roottestclusterallreplicas-gw4_default Creating Stderr: Network roottestclusterallreplicas-gw4_default Created Stderr: Container roottestclusterallreplicas-gw4-node2-1 Creating Stderr: Container roottestclusterallreplicas-gw4-node1-1 Creating Stderr: Container roottestclusterallreplicas-gw4-node2-1 Created Stderr: Container roottestclusterallreplicas-gw4-node1-1 Created Stderr: Container roottestclusterallreplicas-gw4-node2-1 Starting Stderr: Container roottestclusterallreplicas-gw4-node1-1 Starting Stderr: Container roottestclusterallreplicas-gw4-node2-1 Started Stderr: Container roottestclusterallreplicas-gw4-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.2.3... http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/.env --project-name roottestbackuprestore-gw6 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/.env --project-name roottestbackuprestore-gw6 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/node/docker-compose.yml up -d --no-recreate] http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stderr: Network roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3_default Creating Stderr: Network roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3_default Created Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1 Creating Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1 Created Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1 Starting Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1 Started ClickHouse instance created get_instance_ip instance_name=local_node http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=local_node http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in local_node, ip: 172.16.5.2... http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None Stderr: Network roottestasyncloaddatabases-gw9_default Creating Stderr: Network roottestasyncloaddatabases-gw9_default Created http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Stderr: Container roottestasyncloaddatabases-gw9-node2-1 Creating Stderr: Container roottestasyncloaddatabases-gw9-node1-1 Creating Stderr: Container roottestasyncloaddatabases-gw9-node2-1 Created Stderr: Container roottestasyncloaddatabases-gw9-node1-1 Created Stderr: Container roottestasyncloaddatabases-gw9-node2-1 Starting Stderr: Container roottestasyncloaddatabases-gw9-node1-1 Starting Stderr: Container roottestasyncloaddatabases-gw9-node2-1 Started Stderr: Container roottestasyncloaddatabases-gw9-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw9-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw9-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.3.3... http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw9-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None Stderr: Network roottestallowfeaturetier-gw8_default Creating Stderr: Network roottestallowfeaturetier-gw8_default Created Stderr: Container roottestallowfeaturetier-gw8-instance-1 Creating Stderr: Container roottestallowfeaturetier-gw8-instance-1 Created Stderr: Container roottestallowfeaturetier-gw8-instance-1 Starting Stderr: Container roottestallowfeaturetier-gw8-instance-1 Started ClickHouse instance created get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestallowfeaturetier-gw8-instance-1/json HTTP/1.1" 200 None get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestallowfeaturetier-gw8-instance-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in instance, ip: 172.16.4.2... http://localhost:None "GET /v1.46/containers/roottestallowfeaturetier-gw8-instance-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None Stderr:time="2025-04-02T02:17:17Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestconfigsubstitutions-gw7_default Creating Stderr: Network roottestconfigsubstitutions-gw7_default Created Stderr: Container roottestconfigsubstitutions-gw7-zoo2-1 Creating Stderr: Container roottestconfigsubstitutions-gw7-zoo3-1 Creating Stderr: Container roottestconfigsubstitutions-gw7-zoo1-1 Creating Stderr: Container roottestconfigsubstitutions-gw7-zoo3-1 Created Stderr: Container roottestconfigsubstitutions-gw7-zoo1-1 Created Stderr: Container roottestconfigsubstitutions-gw7-zoo2-1 Created Stderr: Container roottestconfigsubstitutions-gw7-zoo2-1 Starting Stderr: Container roottestconfigsubstitutions-gw7-zoo3-1 Starting Stderr: Container roottestconfigsubstitutions-gw7-zoo1-1 Starting Stderr: Container roottestconfigsubstitutions-gw7-zoo2-1 Started Stderr: Container roottestconfigsubstitutions-gw7-zoo1-1 Started Stderr: Container roottestconfigsubstitutions-gw7-zoo3-1 Started Stderr:time="2025-04-02T02:17:19Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:17:19Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.6.2, port:2181, use_ssl:False Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None Stderr: Network roottestbackuprestore-gw6_default Creating Stderr: Network roottestbackuprestore-gw6_default Created Stderr: Container roottestbackuprestore-gw6-node-1 Creating Stderr: Container roottestbackuprestore-gw6-node-1 Created Stderr: Container roottestbackuprestore-gw6-node-1 Starting Stderr: Container roottestbackuprestore-gw6-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestbackuprestore-gw6-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestbackuprestore-gw6-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.8.2... http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestbackuprestore-gw6-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Stderr:time="2025-04-02T02:17:17Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestdatabasebackup-gw1_default Creating Stderr: Network roottestdatabasebackup-gw1_default Created Stderr: Volume "roottestdatabasebackup-gw1_data1-1" Creating Stderr: Volume "roottestdatabasebackup-gw1_data1-1" Created Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Creating Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Creating Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Created Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Created Stderr: Container roottestdatabasebackup-gw1-minio1-1 Creating Stderr: Container roottestdatabasebackup-gw1-resolver-1 Creating Stderr: Container roottestdatabasebackup-gw1-resolver-1 Created Stderr: Container roottestdatabasebackup-gw1-minio1-1 Created Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Starting Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Starting Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Started Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Started Stderr: Container roottestdatabasebackup-gw1-minio1-1 Starting Stderr: Container roottestdatabasebackup-gw1-resolver-1 Starting Stderr: Container roottestdatabasebackup-gw1-minio1-1 Started Stderr: Container roottestdatabasebackup-gw1-resolver-1 Started Stderr:time="2025-04-02T02:17:19Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:17:19Z" level=debug msg="otel error" error="" Trying to connect to Minio... get_instance_ip instance_name=minio1 http://localhost:None "GET /v1.46/containers/roottestdatabasebackup-gw1-minio1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=proxy1 http://localhost:None "GET /v1.46/containers/roottestdatabasebackup-gw1-proxy1-1/json HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.7.4:9001 Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (2): 172.16.7.4:9001 Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (3): 172.16.7.4:9001 Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (4): 172.16.7.4:9001 Can't connect to Minio: HTTPConnectionPool(host='172.16.7.4', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e81602a220c39d6378f3ea4ab85303ecee230d5493133567cdf51e2196f44786/json HTTP/1.1" 200 None ClickHouse local_node started Executing query CREATE DATABASE IF NOT EXISTS test on local_node http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None Stderr: node Pulling Stderr: redis1 Pulling Stderr: 6e909acdb790 Pulling fs layer Stderr: d9b98f603d6b Pulling fs layer Stderr: 513666b05c81 Pulling fs layer Stderr: 764e7fbfdb05 Pulling fs layer Stderr: 2884ccae34c1 Pulling fs layer Stderr: df4660a56a60 Pulling fs layer Stderr: 4f4fb700ef54 Pulling fs layer Stderr: b138a3179b82 Pulling fs layer Stderr: 6e909acdb790 Waiting Stderr: d9b98f603d6b Waiting Stderr: df4660a56a60 Waiting Stderr: 2884ccae34c1 Waiting Stderr: 513666b05c81 Waiting Stderr: 764e7fbfdb05 Waiting Stderr: 4f4fb700ef54 Waiting Stderr: b138a3179b82 Waiting Stderr: node Pulled Stderr: 6e909acdb790 Downloading [> ] 299.9kB/28.2MB Stderr: d9b98f603d6b Download complete Stderr: 513666b05c81 Downloading [==================================================>] 873B/873B Stderr: 513666b05c81 Verifying Checksum Stderr: 513666b05c81 Download complete Stderr: 764e7fbfdb05 Downloading [> ] 22.47kB/1.438MB Stderr: 764e7fbfdb05 Download complete Stderr: 6e909acdb790 Verifying Checksum Stderr: 6e909acdb790 Download complete Stderr: 2884ccae34c1 Downloading [> ] 160.4kB/15.33MB Stderr: df4660a56a60 Downloading [==================================================>] 97B/97B Stderr: df4660a56a60 Verifying Checksum Stderr: df4660a56a60 Download complete Stderr: 4f4fb700ef54 Downloading [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Verifying Checksum Stderr: 4f4fb700ef54 Download complete Stderr: b138a3179b82 Downloading [==================================================>] 572B/572B Stderr: b138a3179b82 Verifying Checksum Stderr: b138a3179b82 Download complete Stderr: 2884ccae34c1 Verifying Checksum Stderr: 2884ccae34c1 Download complete Stderr: 6e909acdb790 Extracting [> ] 294.9kB/28.2MB Stderr: 6e909acdb790 Extracting [=====> ] 3.244MB/28.2MB Stderr: 6e909acdb790 Extracting [===========> ] 6.488MB/28.2MB Stderr: 6e909acdb790 Extracting [================> ] 9.142MB/28.2MB Stderr: 6e909acdb790 Extracting [==================> ] 10.62MB/28.2MB Stderr: 6e909acdb790 Extracting [======================> ] 12.68MB/28.2MB Stderr: 6e909acdb790 Extracting [==============================> ] 17.4MB/28.2MB Stderr: 6e909acdb790 Extracting [=====================================> ] 21.23MB/28.2MB Stderr: 6e909acdb790 Extracting [=========================================> ] 23.3MB/28.2MB Stderr: 6e909acdb790 Extracting [==========================================> ] 23.89MB/28.2MB Stderr: 6e909acdb790 Extracting [===============================================> ] 26.54MB/28.2MB Stderr: 6e909acdb790 Extracting [================================================> ] 27.13MB/28.2MB Stderr: 6e909acdb790 Extracting [================================================> ] 27.43MB/28.2MB Stderr: 6e909acdb790 Extracting [==================================================>] 28.2MB/28.2MB Stderr: 6e909acdb790 Pull complete Stderr: d9b98f603d6b Extracting [==================================================>] 1.102kB/1.102kB Stderr: d9b98f603d6b Extracting [==================================================>] 1.102kB/1.102kB Stderr: d9b98f603d6b Pull complete Stderr: 513666b05c81 Extracting [==================================================>] 873B/873B Stderr: 513666b05c81 Extracting [==================================================>] 873B/873B Stderr: 513666b05c81 Pull complete Stderr: 764e7fbfdb05 Extracting [=> ] 32.77kB/1.438MB Stderr: 764e7fbfdb05 Extracting [==================================================>] 1.438MB/1.438MB Stderr: 764e7fbfdb05 Extracting [==================================================>] 1.438MB/1.438MB Stderr: 764e7fbfdb05 Pull complete Stderr: 2884ccae34c1 Extracting [> ] 163.8kB/15.33MB Stderr: 2884ccae34c1 Extracting [=============> ] 4.096MB/15.33MB Stderr: 2884ccae34c1 Extracting [========================> ] 7.373MB/15.33MB Stderr: 2884ccae34c1 Extracting [===================================> ] 10.98MB/15.33MB Stderr: 2884ccae34c1 Extracting [=================================================> ] 15.24MB/15.33MB Stderr: 2884ccae34c1 Extracting [==================================================>] 15.33MB/15.33MB Stderr: 2884ccae34c1 Pull complete Stderr: df4660a56a60 Extracting [==================================================>] 97B/97B Stderr: df4660a56a60 Extracting [==================================================>] 97B/97B Stderr: df4660a56a60 Pull complete http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Pull complete Stderr: b138a3179b82 Extracting [==================================================>] 572B/572B Stderr: b138a3179b82 Extracting [==================================================>] 572B/572B Stderr: b138a3179b82 Pull complete Stderr: redis1 Pulled Setup Redis Command:[docker compose --project-name roottestdictionariesredis-gw2 --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml --verbose up -d] http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/b47299568d4d8ada2992bedccbe853616ad568090895d276ff64400340125b05/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw9-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw9-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.3.2... http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw9-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f692a5edc1d68ceffbb44f4fecac0567e13e20b4b57451aa6c665fa41e78fab7/json HTTP/1.1" 200 None ClickHouse node2 started Executing query CREATE DATABASE IF NOT EXISTS dict ENGINE=Dictionary; CREATE DATABASE IF NOT EXISTS test; on node1 http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query CREATE TABLE test.table_LocalClickHouse_flat_ ( KeyField UInt64,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64,ParentKeyField UInt64) ENGINE MergeTree ORDER BY tuple(); on local_node http://localhost:None "GET /v1.46/containers/1f325248640fc8370a409b66d840b666675223cd4412a349ff84787a7c817e1f/json HTTP/1.1" 200 None ClickHouse instance started Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None Executing query CREATE DATABASE IF NOT EXISTS dict ENGINE=Dictionary; CREATE DATABASE IF NOT EXISTS test; on node2 Executing query INSERT INTO test.table_LocalClickHouse_flat_ (KeyField,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_,ParentKeyField) values (1,22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4,0),(2,3,4,5,6,-7,-8,-9,-10,'550e8400-e29b-41d4-a716-446655440002','1978-06-28','1986-02-28 23:42:25','hello',21.543,3222154213.4,1) on local_node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=1 on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None Executing query system flush logs on node2 Starting new HTTP connection (5): 172.16.7.4:9001 http://172.16.7.4:9001 "GET / HTTP/1.1" 200 0 Connected to Minio. http://172.16.7.4:9001 "GET /root?location= HTTP/1.1" 404 0 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None http://172.16.7.4:9001 "PUT /root HTTP/1.1" 200 0 S3 bucket 'root' created http://172.16.7.4:9001 "GET /root2?location= HTTP/1.1" 404 0 http://172.16.7.4:9001 "PUT /root2 HTTP/1.1" 200 0 S3 bucket 'root2' created ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --project-name roottestdatabasebackup-gw1 --file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --project-name roottestdatabasebackup-gw1 --file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate] Executing query CREATE DATABASE IF NOT EXISTS test on local_node Stderr:time="2025-04-02T02:17:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestdictionariesredis-gw2_default Creating Stderr: Network roottestdictionariesredis-gw2_default Created Stderr: Container roottestdictionariesredis-gw2-redis1-1 Creating Stderr: Container roottestdictionariesredis-gw2-redis1-1 Created Stderr: Container roottestdictionariesredis-gw2-redis1-1 Starting Stderr: Container roottestdictionariesredis-gw2-redis1-1 Started Stderr:time="2025-04-02T02:17:21Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:17:21Z" level=debug msg="otel error" error="" http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] http://localhost:None "GET /v1.46/containers/0ace7adff2ab0e7efc0e97e613291501e7430642c5b8bc5022712e5bede9dbbc/json HTTP/1.1" 200 None ClickHouse node started Executing query CREATE DATABASE test ENGINE = Ordinary on node Executing query CREATE TABLE test.table_LocalClickHouse_hashed_ ( KeyField UInt64,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64,ParentKeyField UInt64) ENGINE MergeTree ORDER BY tuple(); on local_node Executing query SYSTEM RELOAD CONFIG on instance Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.1.3, port:2181, use_ssl:False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.1.2, port:2181, use_ssl:False Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Running Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Running Stderr: Container roottestdatabasebackup-gw1-minio1-1 Running Stderr: Container roottestdatabasebackup-gw1-instance-1 Creating Stderr: Container roottestdatabasebackup-gw1-resolver-1 Running Stderr: Container roottestdatabasebackup-gw1-instance-1 Created Stderr: Container roottestdatabasebackup-gw1-instance-1 Starting Stderr: Container roottestdatabasebackup-gw1-instance-1 Started ClickHouse instance created get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestdatabasebackup-gw1-instance-1/json HTTP/1.1" 200 None get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestdatabasebackup-gw1-instance-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in instance, ip: 172.16.7.6... http://localhost:None "GET /v1.46/containers/roottestdatabasebackup-gw1-instance-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --project-name roottestbackuprestoreoncluster-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --project-name roottestbackuprestoreoncluster-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/docker-compose.yml up -d --no-recreate] Executing query INSERT INTO test.table_LocalClickHouse_hashed_ (KeyField,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_,ParentKeyField) values (1,22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4,0),(2,3,4,5,6,-7,-8,-9,-10,'550e8400-e29b-41d4-a716-446655440002','1978-06-28','1986-02-28 23:42:25','hello',21.543,3222154213.4,1) on local_node http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS test.tbl on node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None Executing query CREATE DATABASE IF NOT EXISTS test on local_node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query CREATE TABLE test.tbl (p Date, k Int8) ENGINE = MergeTree PARTITION BY toYYYYMM(p) ORDER BY p on node Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=1 on instance http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None Executing query CREATE TABLE test.table_LocalClickHouse_cache_ ( KeyField UInt64,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64,ParentKeyField UInt64) ENGINE MergeTree ORDER BY tuple(); on local_node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(1), 1) on node http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query SELECT 1 SETTINGS enable_parallel_replicas=1 on instance http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Running Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Running Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Running Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Started Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Started Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.1.5... http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(2), 2) on node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None Executing query INSERT INTO test.table_LocalClickHouse_cache_ (KeyField,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_,ParentKeyField) values (1,22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4,0),(2,3,4,5,6,-7,-8,-9,-10,'550e8400-e29b-41d4-a716-446655440002','1978-06-28','1986-02-28 23:42:25','hello',21.543,3222154213.4,1) on local_node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/2/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/2/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query rename table system.text_log to system.text_log_1_test on node2 http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None Executing query SYSTEM RELOAD CONFIG on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(3), 3) on node http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None Executing query CREATE DATABASE IF NOT EXISTS test on local_node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None Executing query rename table system.query_log to system.query_log_1_test on node2 http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None Executing query CREATE TABLE test.table_LocalClickHouse_direct_ ( KeyField UInt64,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64,ParentKeyField UInt64) ENGINE MergeTree ORDER BY tuple(); on local_node Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(31), 31) on node Executing query system flush logs on node2 http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=1 on instance http://localhost:None "GET /v1.46/containers/111889b494951dc196c8eee2269adae045f2d44d534620bfcb894ee37efb4edd/json HTTP/1.1" 200 None ClickHouse instance started run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query INSERT INTO test.table_LocalClickHouse_direct_ (KeyField,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_,ParentKeyField) values (1,22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4,0),(2,3,4,5,6,-7,-8,-9,-10,'550e8400-e29b-41d4-a716-446655440002','1978-06-28','1986-02-28 23:42:25','hello',21.543,3222154213.4,1) on local_node http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_database_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table_1 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_1 VALUES (0, 'test_database.test_table_1'); CREATE TABLE test_database.test_table_2 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_2 VALUES (0, 'test_database.test_table_2'); CREATE TABLE test_database.test_table_3 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_3 VALUES (0, 'test_database.test_table_3'); BACKUP DATABASE test_database TO Disk('backup_disk_local', 'test_database_backup'); CREATE DATABASE test_database_backup ENGINE = Backup('test_database', Disk('backup_disk_local', 'test_database_backup')); on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(32), 32) on node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None Executing query SELECT 1 SETTINGS enable_parallel_replicas=1 on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None Executing query CREATE DATABASE IF NOT EXISTS test on local_node Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(33), 33) on node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/2/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/2/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query SYSTEM RELOAD CONFIG on instance Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None Executing query CREATE TABLE test.table_LocalClickHouse_complex_key_hashed_ ( KeyField1 UInt64,KeyField2 String,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64) ENGINE MergeTree ORDER BY tuple(); on local_node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query SELECT * FROM test.tbl ORDER BY p on node http://localhost:None "GET /v1.46/containers/0836a62fbdf9a82494b60b1cfab27a6e7a93eb990edac2e75243dfe22d121621/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.1.7... http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3180ab299bc4c56446ee2fb95783c886cc55e40315c0a4f8d8f87c5681b05ba1/json HTTP/1.1" 200 None Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3180ab299bc4c56446ee2fb95783c886cc55e40315c0a4f8d8f87c5681b05ba1/json HTTP/1.1" 200 None Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query INSERT INTO test.table_LocalClickHouse_complex_key_hashed_ (KeyField1,KeyField2,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_) values (1,'world',22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4),(2,'qwerty2',52,2345,6544,9191991,-2,-717,-81818,-92929,'550e8400-e29b-41d4-a716-446655440007','1975-09-28','2000-02-28 23:33:24','my',255.543,3332221.44) on local_node Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3180ab299bc4c56446ee2fb95783c886cc55e40315c0a4f8d8f87c5681b05ba1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3180ab299bc4c56446ee2fb95783c886cc55e40315c0a4f8d8f87c5681b05ba1/json HTTP/1.1" 200 None ClickHouse node2 started get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node3, ip: 172.16.1.6... Executing query rename table system.text_log to system.text_log_2_test on node2 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f93869b969fdc376bbf8097d5845fcfae84d9cff7a73cd1539dfca7a60988a0c/json HTTP/1.1" 200 None ClickHouse node3 started Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query ALTER TABLE test.tbl FREEZE on node [gw8] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query CREATE DATABASE IF NOT EXISTS test on local_node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query rename table system.query_log to system.query_log_2_test on node2 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query CREATE TABLE test.table_LocalClickHouse_complex_key_cache_ ( KeyField1 UInt64,KeyField2 String,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64) ENGINE MergeTree ORDER BY tuple(); on local_node Executing query CREATE TABLE test.tbl2 AS test.tbl on node Executing query DROP TABLE IF EXISTS test_experimental on instance run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:02 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] Executing query INSERT INTO tbl VALUES (1) on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Executing query system flush logs on node2 Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query INSERT INTO test.table_LocalClickHouse_complex_key_cache_ (KeyField1,KeyField2,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_) values (1,'world',22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4),(2,'qwerty2',52,2345,6544,9191991,-2,-717,-81818,-92929,'550e8400-e29b-41d4-a716-446655440007','1975-09-28','2000-02-28 23:33:24','my',255.543,3332221.44) on local_node Stdout:8 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(3), 3) on node Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.6.3, port:2181, use_ssl:False http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Connecting to 172.16.6.3(172.16.6.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query SYSTEM RELOAD CONFIG on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.6.4, port:2181, use_ssl:False Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query BACKUP TABLE tbl TO Disk('backups', '1') ASYNC on node1 via HTTP interface Starting new HTTP connection (1): 172.16.1.5:8123 http://172.16.1.5:8123 "GET /?query=BACKUP+TABLE+tbl++TO+Disk%28%27backups%27%2C+%271%27%29+ASYNC HTTP/1.1" 200 None Executing query CREATE DATABASE IF NOT EXISTS test on local_node Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(4), 4) on node Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.6.2, port:2181, use_ssl:False Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query BACKUP TABLE tbl TO Disk('backups', '1') ASYNC on node2 via HTTP interface Starting new HTTP connection (1): 172.16.1.7:8123 Zookeeper connection established, state: CONNECTED run_kazoo_commands_with_retries: zoo1, .create_zk_roots at 0x7fa8749b8550> Sending request(xid=1): Create(path='/setting/max_query_size', data=b'77777', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received error(xid=1) NoNodeError() Sending request(xid=2): Exists(path='/setting', watcher=None) Sending request(xid=3): Exists(path='/', watcher=None) http://172.16.1.7:8123 "GET /?query=BACKUP+TABLE+tbl++TO+Disk%28%27backups%27%2C+%271%27%29+ASYNC HTTP/1.1" 200 None Received response(xid=3): ZnodeStat(czxid=0, mzxid=0, ctime=0, mtime=0, version=0, cversion=0, aversion=0, ephemeralOwner=0, dataLength=0, numChildren=1, pzxid=0) Sending request(xid=4): Create(path='/setting', data=b'', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Executing query SELECT status FROM system.backups WHERE id IN ['f2de5b9f-7d74-421c-b4c0-1efd351e6986', '4cf586c8-f0ab-4b70-b21f-901f1382f828'] AND status == 'CREATING_BACKUP' on node1 Received response(xid=4): '/setting' Sending request(xid=5): Create(path='/setting/max_query_size', data=b'77777', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=5): '/setting/max_query_size' Sending request(xid=6): Create(path='/users_from_zk_1', data=b'default', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=6): '/users_from_zk_1' Sending request(xid=7): Create(path='/users_from_zk_2', data=b'default', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=7): '/users_from_zk_2' Sending request(xid=8): Create(path='/min_bytes_for_wide_part', data=b'33', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=8): '/min_bytes_for_wide_part' Sending request(xid=9): Create(path='/merge_max_block_size', data=b'8888', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=9): '/merge_max_block_size' ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/.env --project-name roottestconfigsubstitutions-gw7 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node8/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/.env --project-name roottestconfigsubstitutions-gw7 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node8/docker-compose.yml up -d --no-recreate] Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query CREATE TABLE test.table_LocalClickHouse_complex_key_direct_ ( KeyField1 UInt64,KeyField2 String,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64) ENGINE MergeTree ORDER BY tuple(); on local_node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(33), 33) on node Executing query SELECT status FROM system.backups WHERE id IN ['f2de5b9f-7d74-421c-b4c0-1efd351e6986', '4cf586c8-f0ab-4b70-b21f-901f1382f828'] AND status == 'CREATING_BACKUP' on node2 Executing query CREATE TABLE test_experimental (uid String, version UInt32, is_deleted UInt8) ENGINE = ReplacingMergeTree(version, is_deleted) ORDER by (uid) SETTINGS allow_experimental_replacing_merge_with_cleanup=1; on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query INSERT INTO test.table_LocalClickHouse_complex_key_direct_ (KeyField1,KeyField2,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_) values (1,'world',22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4),(2,'qwerty2',52,2345,6544,9191991,-2,-717,-81818,-92929,'550e8400-e29b-41d4-a716-446655440007','1975-09-28','2000-02-28 23:33:24','my',255.543,3332221.44) on local_node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(34), 34) on node Executing query SELECT count() FROM system.backups WHERE id IN ['f2de5b9f-7d74-421c-b4c0-1efd351e6986', '4cf586c8-f0ab-4b70-b21f-901f1382f828'] AND status == 'BACKUP_CREATED' on node1 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query CREATE DATABASE IF NOT EXISTS test on local_node Executing query SYSTEM RELOAD CONFIG on instance Executing query rename table system.text_log to system.text_log_3_test on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query SELECT * FROM test.tbl2 ORDER BY p on node Stdout:8 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query SELECT count() FROM system.backups WHERE id IN ['f2de5b9f-7d74-421c-b4c0-1efd351e6986', '4cf586c8-f0ab-4b70-b21f-901f1382f828'] AND status == 'BACKUP_CREATED' on node2 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query rename table system.query_log to system.query_log_3_test on node2 Executing query CREATE TABLE test.table_LocalClickHouse_range_hashed_ ( KeyField1 UInt64,KeyField2 Date,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64,StartDate Date,EndDate Date) ENGINE MergeTree ORDER BY tuple(); on local_node Stderr: Container roottestconfigsubstitutions-gw7-node2-1 Creating Stderr: Container roottestconfigsubstitutions-gw7-zoo3-1 Running Stderr: Container roottestconfigsubstitutions-gw7-node1-1 Creating Stderr: Container roottestconfigsubstitutions-gw7-node5-1 Creating Stderr: Container roottestconfigsubstitutions-gw7-node8-1 Creating Stderr: Container roottestconfigsubstitutions-gw7-node6-1 Creating Stderr: Container roottestconfigsubstitutions-gw7-zoo1-1 Running Stderr: Container roottestconfigsubstitutions-gw7-node7-1 Creating Stderr: Container roottestconfigsubstitutions-gw7-zoo2-1 Running Stderr: Container roottestconfigsubstitutions-gw7-node4-1 Creating Stderr: Container roottestconfigsubstitutions-gw7-node3-1 Creating Stderr: Container roottestconfigsubstitutions-gw7-node1-1 Created Stderr: Container roottestconfigsubstitutions-gw7-node8-1 Created Stderr: Container roottestconfigsubstitutions-gw7-node3-1 Created Stderr: Container roottestconfigsubstitutions-gw7-node2-1 Created Stderr: Container roottestconfigsubstitutions-gw7-node5-1 Created Stderr: Container roottestconfigsubstitutions-gw7-node7-1 Created Stderr: Container roottestconfigsubstitutions-gw7-node6-1 Created Stderr: Container roottestconfigsubstitutions-gw7-node4-1 Created Stderr: Container roottestconfigsubstitutions-gw7-node4-1 Starting Stderr: Container roottestconfigsubstitutions-gw7-node8-1 Starting Stderr: Container roottestconfigsubstitutions-gw7-node7-1 Starting Stderr: Container roottestconfigsubstitutions-gw7-node6-1 Starting Stderr: Container roottestconfigsubstitutions-gw7-node2-1 Starting Stderr: Container roottestconfigsubstitutions-gw7-node1-1 Starting Stderr: Container roottestconfigsubstitutions-gw7-node3-1 Starting Stderr: Container roottestconfigsubstitutions-gw7-node5-1 Starting Stderr: Container roottestconfigsubstitutions-gw7-node3-1 Started Stderr: Container roottestconfigsubstitutions-gw7-node5-1 Started Stderr: Container roottestconfigsubstitutions-gw7-node8-1 Started Stderr: Container roottestconfigsubstitutions-gw7-node4-1 Started Stderr: Container roottestconfigsubstitutions-gw7-node6-1 Started Stderr: Container roottestconfigsubstitutions-gw7-node2-1 Started Stderr: Container roottestconfigsubstitutions-gw7-node7-1 Started Stderr: Container roottestconfigsubstitutions-gw7-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.6.12... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query SELECT count() FROM system.backups WHERE id IN ['f2de5b9f-7d74-421c-b4c0-1efd351e6986', '4cf586c8-f0ab-4b70-b21f-901f1382f828'] AND status == 'BACKUP_FAILED' on node1 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None copy from /var/lib/clickhouse/shadow/*/data/test/tbl to /var/lib/clickhouse/data/test/tbl2/detached run container_id:roottestbackuprestore-gw6-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl2/detached'] Command:[docker exec roottestbackuprestore-gw6-node-1 bash -c cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl2/detached] http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None Executing query ALTER TABLE test.tbl2 ATTACH PARTITION 197001 on node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query system flush logs on node2 http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query INSERT INTO test.table_LocalClickHouse_range_hashed_ (KeyField1,StartDate,EndDate,KeyField2,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_) values (1,'2019-02-01','2019-02-28','2019-02-10',22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4),(2,'2019-04-01','2019-04-28','2019-04-10',11,3223,41444,52515,-65,-747,-8388,-9099,'550e8400-e29b-41d4-a716-446655440004','1973-06-29','2002-02-28 23:23:25','!!!!',32.543,3332543.4) on local_node Executing query CREATE TABLE test_experimental (uid String, version UInt32, is_deleted UInt8) ENGINE = ReplacingMergeTree(version, is_deleted) ORDER by (uid) SETTINGS allow_experimental_replacing_merge_with_cleanup=1; on instance run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT count() FROM system.backups WHERE id IN ['f2de5b9f-7d74-421c-b4c0-1efd351e6986', '4cf586c8-f0ab-4b70-b21f-901f1382f828'] AND status == 'BACKUP_FAILED' on node2 http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Stdout:8 Executing query ALTER TABLE test.tbl2 ATTACH PARTITION 197002 on node http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query system reload dictionaries on local_node http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None Executing query SELECT engine_full FROM system.tables WHERE name = 'test_experimental' on instance Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query SELECT sleep(2) on node http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None Executing query SYSTEM RELOAD CONFIG on instance http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query RESTORE TABLE tbl FROM Disk('backups', '1') on node1 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query select dictGetUInt8('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query rename table system.text_log to system.text_log_4_test on node2 http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestallowfeaturetier-gw8-instance-1 bash -c ps -C clickhouse] Stdout:8 Stdout:758 Executing query SELECT * FROM tbl on node1 http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:02 clickhouse run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestallowfeaturetier-gw8-instance-1 bash -c pkill clickhouse] http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query rename table system.query_log to system.query_log_4_test on node2 http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 Stdout:8 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/795e3f0c0fc048f8696b0a9728826b54fc91ba56f2b0c35f343c1085c1996168/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.6.10... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ea181959a21a4952f1bac603a4fb36c2aa1ef110e821b36cfa20ece4a1a072ec/json HTTP/1.1" 200 None ClickHouse node2 started get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node3, ip: 172.16.6.5... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/716c068ddaa6efae9ad1e636e597e50f6be5d94cc5090489e62eda34234d2b1b/json HTTP/1.1" 200 None ClickHouse node3 started get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node4-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node4-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node4, ip: 172.16.6.9... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node4-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/91a505f90ed9163e6771ccd62e75095f28193dfb1e04ebad5f9bad8deba20191/json HTTP/1.1" 200 None ClickHouse node4 started get_instance_ip instance_name=node5 Executing query system flush logs on node2 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node5-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node5 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node5-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node5, ip: 172.16.6.7... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node5-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/54b7cc523e9cd447d0ccd355c36b1fdb040d41504927a29eb6ffc3f8653bade7/json HTTP/1.1" 200 None ClickHouse node5 started get_instance_ip instance_name=node6 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node6-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node6 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node6-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node6, ip: 172.16.6.11... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node6-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e4f849560eeeb44fea1d321d9cf9e7e0c07fcb5df96ffa27360642ad46032c09/json HTTP/1.1" 200 None ClickHouse node6 started get_instance_ip instance_name=node7 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node7-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node7 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node7-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node7, ip: 172.16.6.8... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node7-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bb61b1652d27fe216584975ac4d88deea4cbee2d9c78a459a083dc6b9a717d47/json HTTP/1.1" 200 None ClickHouse node7 started get_instance_ip instance_name=node8 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node8-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node8 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node8-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node8, ip: 172.16.6.6... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-node8-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9115ffa0ae82c6d504faf365acbbaa5d4781269bf92c947f1b985610f01e1312/json HTTP/1.1" 200 None ClickHouse node8 started Executing query DROP DATABASE IF EXISTS db1 SYNC on node5 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query CREATE DATABASE db1 on node5 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query CREATE TABLE db1.test_table(date Date, k1 String, v1 Int32) ENGINE = MergeTree(date, (k1, date), 8192) on node5 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt16('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "POST /v1.46/exec/eaccc76c6dc4020cb2d3013cd5e2cc56b14eedfd8a1610f5fe1ff6063e26be46/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/eaccc76c6dc4020cb2d3013cd5e2cc56b14eedfd8a1610f5fe1ff6063e26be46/json HTTP/1.1" 200 586 Executing query INSERT INTO db1.test_table VALUES('2000-01-01', 'test_key', 1) on node5 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query rename table system.text_log to system.text_log_5_test on node2 Stdout:8 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query SELECT * FROM test.tbl2 ORDER BY p on node Executing query SELECT name FROM system.databases WHERE name = 'db1' on node5 Executing query rename table system.query_log to system.query_log_5_test on node2 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query ALTER TABLE test.tbl2 UPDATE k=10 WHERE 1 on node Executing query system flush logs on node2 Executing query SELECT name FROM system.tables WHERE database = 'db1' AND name = 'test_table' on node5 http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None Executing query SELECT sleep(2) on node Executing query SELECT name FROM system.columns WHERE database = 'db1' AND table = 'test_table' on node5 Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] http://localhost:None "GET /v1.46/containers/326759a587bc0f65968dc3f1ab94df373692ec32452afb73d73baecf68389c6e/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.2.2... http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node2-1/json HTTP/1.1" 200 None Stderr: mysql80 Pulling Stderr: node Pulling Stderr: cea172a6e83b Pulling fs layer Stderr: 28e01aa53f13 Pulling fs layer Stderr: e5fa3211d7a7 Pulling fs layer Stderr: 753b8441f7e6 Pulling fs layer Stderr: b1339a14fa1a Pulling fs layer Stderr: be386ff914e3 Pulling fs layer Stderr: 93272c957f26 Pulling fs layer Stderr: c106a4902288 Pulling fs layer Stderr: 036f4325df2d Pulling fs layer Stderr: 1d34979e7120 Pulling fs layer Stderr: de67a2f637e5 Pulling fs layer Stderr: be386ff914e3 Waiting Stderr: 1d34979e7120 Waiting Stderr: c106a4902288 Waiting Stderr: de67a2f637e5 Waiting http://localhost:None "GET /v1.46/containers/a54f5e967d362c32ae36bda7afaa68597bf845ecaa0043a0457557f2b3b8b53b/json HTTP/1.1" 200 None Stderr: 93272c957f26 Waiting Stderr: 753b8441f7e6 Waiting Stderr: b1339a14fa1a Waiting Stderr: cea172a6e83b Downloading [> ] 511.4kB/49.09MB Stderr: e5fa3211d7a7 Downloading [> ] 13.78kB/983kB Stderr: 28e01aa53f13 Downloading [==================================================>] 882B/882B Stderr: 28e01aa53f13 Verifying Checksum Stderr: 28e01aa53f13 Download complete Stderr: e5fa3211d7a7 Downloading [==================================================>] 983kB/983kB Stderr: e5fa3211d7a7 Verifying Checksum Stderr: e5fa3211d7a7 Download complete Stderr: 753b8441f7e6 Downloading [> ] 92.88kB/6.897MB Stderr: b1339a14fa1a Downloading [==================================================>] 2.606kB/2.606kB Stderr: b1339a14fa1a Verifying Checksum Stderr: b1339a14fa1a Download complete Stderr: be386ff914e3 Downloading [==================================================>] 333B/333B Stderr: be386ff914e3 Verifying Checksum Stderr: be386ff914e3 Download complete Stderr: 93272c957f26 Downloading [> ] 528.7kB/49.62MB Stderr: node Pulled Stderr: 753b8441f7e6 Verifying Checksum Stderr: 753b8441f7e6 Download complete Stderr: c106a4902288 Downloading [==================================================>] 315B/315B Stderr: c106a4902288 Download complete Stderr: 036f4325df2d Downloading [> ] 524.4kB/125.3MB Stderr: cea172a6e83b Downloading [================================> ] 32MB/49.09MB Stderr: 93272c957f26 Downloading [================================> ] 31.85MB/49.62MB Stderr: cea172a6e83b Verifying Checksum Stderr: cea172a6e83b Download complete Stderr: 1d34979e7120 Downloading [==================================> ] 3.647kB/5.327kB Stderr: 1d34979e7120 Downloading [==================================================>] 5.327kB/5.327kB Stderr: 1d34979e7120 Verifying Checksum Stderr: 1d34979e7120 Download complete Stderr: 036f4325df2d Downloading [========> ] 21.98MB/125.3MB Stderr: de67a2f637e5 Downloading [==================================================>] 122B/122B Stderr: de67a2f637e5 Verifying Checksum Stderr: de67a2f637e5 Download complete Stderr: 93272c957f26 Verifying Checksum Stderr: 93272c957f26 Download complete Stderr: cea172a6e83b Extracting [> ] 491.5kB/49.09MB Stderr: 036f4325df2d Downloading [=================> ] 43.71MB/125.3MB Stderr: cea172a6e83b Extracting [=> ] 983kB/49.09MB Stderr: 036f4325df2d Downloading [===========================> ] 69.89MB/125.3MB Stderr: cea172a6e83b Extracting [===> ] 3.441MB/49.09MB Stderr: 036f4325df2d Downloading [=====================================> ] 92.73MB/125.3MB Stderr: cea172a6e83b Extracting [=======> ] 6.881MB/49.09MB Stderr: 036f4325df2d Downloading [===============================================> ] 117.8MB/125.3MB Stderr: 036f4325df2d Verifying Checksum Stderr: 036f4325df2d Download complete Stderr: cea172a6e83b Extracting [=======> ] 7.373MB/49.09MB Stderr: cea172a6e83b Extracting [==========> ] 10.32MB/49.09MB Stderr: cea172a6e83b Extracting [================> ] 15.73MB/49.09MB Stderr: cea172a6e83b Extracting [====================> ] 20.15MB/49.09MB Stderr: cea172a6e83b Extracting [========================> ] 23.59MB/49.09MB Stderr: cea172a6e83b Extracting [============================> ] 27.53MB/49.09MB Stderr: cea172a6e83b Extracting [===============================> ] 30.97MB/49.09MB Stderr: cea172a6e83b Extracting [================================> ] 31.95MB/49.09MB Stderr: cea172a6e83b Extracting [==========================================> ] 41.29MB/49.09MB Stderr: cea172a6e83b Extracting [============================================> ] 43.25MB/49.09MB Stderr: cea172a6e83b Extracting [=============================================> ] 44.73MB/49.09MB Stderr: cea172a6e83b Extracting [==============================================> ] 45.71MB/49.09MB Stderr: cea172a6e83b Extracting [===============================================> ] 46.2MB/49.09MB Stderr: cea172a6e83b Extracting [==================================================>] 49.09MB/49.09MB Stderr: cea172a6e83b Pull complete Stderr: 28e01aa53f13 Extracting [==================================================>] 882B/882B Stderr: 28e01aa53f13 Extracting [==================================================>] 882B/882B Stderr: 28e01aa53f13 Pull complete Stderr: e5fa3211d7a7 Extracting [=> ] 32.77kB/983kB Stderr: e5fa3211d7a7 Extracting [==================================================>] 983kB/983kB Stderr: e5fa3211d7a7 Extracting [==================================================>] 983kB/983kB Stderr: e5fa3211d7a7 Pull complete Stderr: 753b8441f7e6 Extracting [> ] 98.3kB/6.897MB Stderr: 753b8441f7e6 Extracting [=================> ] 2.458MB/6.897MB Stderr: 753b8441f7e6 Extracting [========================> ] 3.441MB/6.897MB Stderr: 753b8441f7e6 Extracting [=================================================> ] 6.881MB/6.897MB Stderr: 753b8441f7e6 Extracting [==================================================>] 6.897MB/6.897MB Stderr: 753b8441f7e6 Pull complete Stderr: b1339a14fa1a Extracting [==================================================>] 2.606kB/2.606kB Stderr: b1339a14fa1a Extracting [==================================================>] 2.606kB/2.606kB Stderr: b1339a14fa1a Pull complete Stderr: be386ff914e3 Extracting [==================================================>] 333B/333B Stderr: be386ff914e3 Extracting [==================================================>] 333B/333B Stderr: be386ff914e3 Pull complete Stderr: 93272c957f26 Extracting [> ] 524.3kB/49.62MB Stderr: 93272c957f26 Extracting [====> ] 4.719MB/49.62MB Stderr: 93272c957f26 Extracting [========> ] 8.913MB/49.62MB Stderr: 93272c957f26 Extracting [=============> ] 13.63MB/49.62MB Stderr: 93272c957f26 Extracting [===================> ] 18.87MB/49.62MB Stderr: 93272c957f26 Extracting [==========================> ] 26.21MB/49.62MB Stderr: 93272c957f26 Extracting [===============================> ] 30.93MB/49.62MB Stderr: 93272c957f26 Extracting [===================================> ] 35.65MB/49.62MB Stderr: 93272c957f26 Extracting [========================================> ] 40.37MB/49.62MB Stderr: 93272c957f26 Extracting [=============================================> ] 45.09MB/49.62MB Stderr: 93272c957f26 Extracting [==================================================>] 49.62MB/49.62MB Stderr: 93272c957f26 Pull complete Stderr: c106a4902288 Extracting [==================================================>] 315B/315B Stderr: c106a4902288 Extracting [==================================================>] 315B/315B Stderr: c106a4902288 Pull complete Stderr: 036f4325df2d Extracting [> ] 557.1kB/125.3MB Stderr: 036f4325df2d Extracting [==> ] 6.685MB/125.3MB Stderr: 036f4325df2d Extracting [====> ] 11.7MB/125.3MB Stderr: 036f4325df2d Extracting [=====> ] 14.48MB/125.3MB Stderr: 036f4325df2d Extracting [=======> ] 18.38MB/125.3MB Stderr: 036f4325df2d Extracting [========> ] 21.17MB/125.3MB Stderr: 036f4325df2d Extracting [========> ] 22.28MB/125.3MB Stderr: 036f4325df2d Extracting [=========> ] 24.51MB/125.3MB Stderr: 036f4325df2d Extracting [===========> ] 27.85MB/125.3MB Stderr: 036f4325df2d Extracting [===========> ] 29.52MB/125.3MB Stderr: 036f4325df2d Extracting [============> ] 31.2MB/125.3MB Stderr: 036f4325df2d Extracting [============> ] 31.75MB/125.3MB Stderr: 036f4325df2d Extracting [============> ] 32.31MB/125.3MB Stderr: 036f4325df2d Extracting [=============> ] 32.87MB/125.3MB Stderr: 036f4325df2d Extracting [=============> ] 33.42MB/125.3MB Stderr: 036f4325df2d Extracting [=============> ] 33.98MB/125.3MB Stderr: 036f4325df2d Extracting [=============> ] 34.54MB/125.3MB Stderr: 036f4325df2d Extracting [==============> ] 35.09MB/125.3MB Stderr: 036f4325df2d Extracting [==============> ] 35.65MB/125.3MB Stderr: 036f4325df2d Extracting [==============> ] 36.21MB/125.3MB Stderr: 036f4325df2d Extracting [==============> ] 36.77MB/125.3MB Stderr: 036f4325df2d Extracting [==============> ] 37.32MB/125.3MB Stderr: 036f4325df2d Extracting [===============> ] 37.88MB/125.3MB Stderr: 036f4325df2d Extracting [===============> ] 38.44MB/125.3MB Stderr: 036f4325df2d Extracting [===============> ] 38.99MB/125.3MB Stderr: 036f4325df2d Extracting [===============> ] 39.55MB/125.3MB Stderr: 036f4325df2d Extracting [================> ] 40.11MB/125.3MB Stderr: 036f4325df2d Extracting [================> ] 40.67MB/125.3MB Stderr: 036f4325df2d Extracting [================> ] 41.22MB/125.3MB Stderr: 036f4325df2d Extracting [================> ] 41.78MB/125.3MB Stderr: 036f4325df2d Extracting [================> ] 42.34MB/125.3MB Stderr: 036f4325df2d Extracting [=================> ] 42.89MB/125.3MB Stderr: 036f4325df2d Extracting [=================> ] 43.45MB/125.3MB Stderr: 036f4325df2d Extracting [=================> ] 44.56MB/125.3MB Stderr: 036f4325df2d Extracting [==================> ] 45.68MB/125.3MB Stderr: 036f4325df2d Extracting [==================> ] 46.24MB/125.3MB Stderr: 036f4325df2d Extracting [==================> ] 46.79MB/125.3MB Stderr: 036f4325df2d Extracting [===================> ] 48.46MB/125.3MB Stderr: 036f4325df2d Extracting [===================> ] 49.58MB/125.3MB Stderr: 036f4325df2d Extracting [=====================> ] 52.92MB/125.3MB Stderr: 036f4325df2d Extracting [======================> ] 56.26MB/125.3MB Stderr: 036f4325df2d Extracting [========================> ] 62.39MB/125.3MB Stderr: 036f4325df2d Extracting [===========================> ] 68.52MB/125.3MB Stderr: 036f4325df2d Extracting [==============================> ] 75.2MB/125.3MB Stderr: 036f4325df2d Extracting [================================> ] 81.89MB/125.3MB Stderr: 036f4325df2d Extracting [==================================> ] 86.34MB/125.3MB Stderr: 036f4325df2d Extracting [====================================> ] 91.91MB/125.3MB Stderr: 036f4325df2d Extracting [=======================================> ] 99.16MB/125.3MB Stderr: 036f4325df2d Extracting [==========================================> ] 107MB/125.3MB Stderr: 036f4325df2d Extracting [==============================================> ] 115.9MB/125.3MB Stderr: 036f4325df2d Extracting [================================================> ] 120.9MB/125.3MB Stderr: 036f4325df2d Extracting [==================================================>] 125.3MB/125.3MB Stderr: 036f4325df2d Pull complete Stderr: 1d34979e7120 Extracting [==================================================>] 5.327kB/5.327kB Stderr: 1d34979e7120 Extracting [==================================================>] 5.327kB/5.327kB Stderr: 1d34979e7120 Pull complete Stderr: de67a2f637e5 Extracting [==================================================>] 122B/122B Stderr: de67a2f637e5 Extracting [==================================================>] 122B/122B Stderr: de67a2f637e5 Pull complete Stderr: mysql80 Pulled Setup MySQL 8 Command:[docker compose --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw5 --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml --verbose up -d] Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/a54f5e967d362c32ae36bda7afaa68597bf845ecaa0043a0457557f2b3b8b53b/json HTTP/1.1" 200 None ClickHouse node2 started Executing query SELECT hostName() FROM clusterAllReplicas('one_shard_two_nodes', system.one) ORDER BY ALL on node1 Stdout:796 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT name FROM system.parts WHERE database = 'db1' AND table = 'test_table' on node5 Stdout:796 Executing query select 20 on instance run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt32('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query INSERT INTO tbl VALUES (1) on node1 No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestallowfeaturetier-gw8-instance-1/exec HTTP/1.1" 201 74 Executing query SELECT hostName(), * FROM clusterAllReplicas("one_shard_two_nodes", system.one) ORDER BY ALL on node1 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 http://localhost:None "POST /v1.46/exec/39f3d830310c7755c50a65d2a11aafa15ccfd477d222191efe40c6ea731ae879/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/39f3d830310c7755c50a65d2a11aafa15ccfd477d222191efe40c6ea731ae879/json HTTP/1.1" 200 586 Executing query SELECT name FROM system.parts_columns WHERE database = 'db1' AND table = 'test_table' on node5 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '2') ASYNC on node1 via HTTP interface Starting new HTTP connection (1): 172.16.1.5:8123 http://172.16.1.5:8123 "GET /?query=BACKUP+TABLE+tbl+ON+CLUSTER+%27cluster%27+TO+Disk%28%27backups%27%2C+%272%27%29+ASYNC HTTP/1.1" 200 None Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '2') ASYNC on node2 via HTTP interface Starting new HTTP connection (1): 172.16.1.7:8123 http://172.16.1.7:8123 "GET /?query=BACKUP+TABLE+tbl+ON+CLUSTER+%27cluster%27+TO+Disk%28%27backups%27%2C+%272%27%29+ASYNC HTTP/1.1" 200 None Executing query SELECT status FROM system.backups WHERE id IN ['89ea04cd-e4c7-43bf-ac1e-4715a3ffd157', 'e28a0bf7-3398-411e-8547-c598d8fbed79'] AND status == 'CREATING_BACKUP' on node1 Executing query SELECT hostName() FROM cluster('one_shard_two_nodes', system.one) on node1 Executing query rename table system.text_log to system.text_log_6_test on node2 Executing query SELECT name FROM system.databases WHERE name = 'db1' on node5 Stderr:time="2025-04-02T02:17:30Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestdictionariesalllayoutsseparatesourcesmysql-gw5_default Creating Stderr: Network roottestdictionariesalllayoutsseparatesourcesmysql-gw5_default Created Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1 Creating Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1 Created Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1 Starting Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1 Started Stderr:time="2025-04-02T02:17:30Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:17:30Z" level=debug msg="otel error" error="" get_instance_ip instance_name=mysql80 http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1/json HTTP/1.1" 200 None Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query rename table system.query_log to system.query_log_6_test on node2 Executing query select 20 on instance Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query SELECT hostName() FROM cluster('one_shard_two_nodes', system.one) on node2 Executing query SELECT name FROM system.tables WHERE database = 'db1' AND name = 'test_table' on node5 ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/.env --project-name roottestdictionariesredis-gw2 --file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/.env --project-name roottestdictionariesredis-gw2 --file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml up -d --no-recreate] Executing query system flush logs on node2 Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query SELECT name FROM system.columns WHERE database = 'db1' AND table = 'test_table' on node5 [gw4] PASSED test_cluster_all_replicas/test.py::test_cluster test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] Executing query SELECT hostName() FROM clusterAllReplicas('one_shard_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=0 on node1 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT status FROM system.backups WHERE id IN ['89ea04cd-e4c7-43bf-ac1e-4715a3ffd157', 'e28a0bf7-3398-411e-8547-c598d8fbed79'] AND status == 'CREATING_BACKUP' on node1 Stdout:777 Clickhouse process running. run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT name FROM system.parts WHERE database = 'db1' AND table = 'test_table' on node5 Stdout:777 Executing query select 20 on instance Executing query select 20 on instance Stderr: Container roottestdictionariesredis-gw2-node-1 Creating Stderr: Container roottestdictionariesredis-gw2-redis1-1 Running Stderr: Container roottestdictionariesredis-gw2-node-1 Created Stderr: Container roottestdictionariesredis-gw2-node-1 Starting Stderr: Container roottestdictionariesredis-gw2-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestdictionariesredis-gw2-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestdictionariesredis-gw2-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.9.3... http://localhost:None "GET /v1.46/containers/roottestdictionariesredis-gw2-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query SELECT status FROM system.backups WHERE id IN ['89ea04cd-e4c7-43bf-ac1e-4715a3ffd157', 'e28a0bf7-3398-411e-8547-c598d8fbed79'] AND status == 'CREATING_BACKUP' on node2 Executing query select dictGetUInt64('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query SELECT name FROM system.parts_columns WHERE database = 'db1' AND table = 'test_table' on node5 http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Executing query SELECT count() FROM system.backups WHERE id IN ['89ea04cd-e4c7-43bf-ac1e-4715a3ffd157', 'e28a0bf7-3398-411e-8547-c598d8fbed79'] AND status == 'BACKUP_CREATED' on node1 http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node [gw7] PASSED test_config_substitutions/test.py::test_allow_databases Executing query select value from system.settings where name = 'max_query_size' on node1 test_config_substitutions/test.py::test_config http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query SELECT * FROM test.tbl2 ORDER BY p on node Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query SELECT count() FROM system.backups WHERE id IN ['89ea04cd-e4c7-43bf-ac1e-4715a3ffd157', 'e28a0bf7-3398-411e-8547-c598d8fbed79'] AND status == 'BACKUP_CREATED' on node2 Executing query select 20 on instance http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query rename table system.text_log to system.text_log_7_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query select value from system.settings where name = 'max_query_size' on node2 http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query DROP TABLE IF EXISTS test.tbl2 on node Executing query SELECT count() FROM system.backups WHERE id IN ['89ea04cd-e4c7-43bf-ac1e-4715a3ffd157', 'e28a0bf7-3398-411e-8547-c598d8fbed79'] AND status == 'BACKUP_FAILED' on node1 http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query rename table system.query_log to system.query_log_7_test on node2 Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query select value from system.settings where name = 'max_query_size' on node3 [gw6] PASSED test_backup_restore/test.py::test_attach_partition Executing query CREATE TABLE test.tbl3 AS test.tbl on node test_backup_restore/test.py::test_replace_partition Executing query DROP DATABASE test_database_backup on instance http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query system flush logs on node2 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT count() FROM system.backups WHERE id IN ['89ea04cd-e4c7-43bf-ac1e-4715a3ffd157', 'e28a0bf7-3398-411e-8547-c598d8fbed79'] AND status == 'BACKUP_FAILED' on node2 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query select value from system.settings where name = 'max_query_size' on node4 Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(3), 3) on node Executing query select 20 on instance http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query DROP DATABASE test_database on instance Executing query select dictGetInt8('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 http://localhost:None "GET /v1.46/containers/1d90b110447d20f902ff71c54cf8bb377bc4bcd183c1f1d8068b2decace8f128/json HTTP/1.1" 200 None ClickHouse node started Preparing RedisSimple_flat_UInt8_ Prepared RedisSimple_flat_UInt8_ Preparing RedisSimple_hashed_UInt8_ Prepared RedisSimple_hashed_UInt8_ Preparing RedisSimple_cache_UInt8_ Prepared RedisSimple_cache_UInt8_ Preparing RedisSimple_direct_UInt8_ Prepared RedisSimple_direct_UInt8_ Preparing RedisHash_complex_key_hashed_UInt8_ Prepared RedisHash_complex_key_hashed_UInt8_ Preparing RedisHash_complex_key_cache_UInt8_ Prepared RedisHash_complex_key_cache_UInt8_ Preparing RedisHash_complex_key_direct_UInt8_ Prepared RedisHash_complex_key_direct_UInt8_ Preparing RedisSimple_flat_UInt16_ Prepared RedisSimple_flat_UInt16_ Preparing RedisSimple_hashed_UInt16_ Prepared RedisSimple_hashed_UInt16_ Preparing RedisSimple_cache_UInt16_ Prepared RedisSimple_cache_UInt16_ Preparing RedisSimple_direct_UInt16_ Prepared RedisSimple_direct_UInt16_ Preparing RedisHash_complex_key_hashed_UInt16_ Prepared RedisHash_complex_key_hashed_UInt16_ Preparing RedisHash_complex_key_cache_UInt16_ Prepared RedisHash_complex_key_cache_UInt16_ Preparing RedisHash_complex_key_direct_UInt16_ Prepared RedisHash_complex_key_direct_UInt16_ Preparing RedisSimple_flat_UInt32_ run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Prepared RedisSimple_flat_UInt32_ Preparing RedisSimple_hashed_UInt32_ Prepared RedisSimple_hashed_UInt32_ Preparing RedisSimple_cache_UInt32_ Executing query select value from system.settings where name = 'max_query_size' on node6 Prepared RedisSimple_cache_UInt32_ Preparing RedisSimple_direct_UInt32_ Prepared RedisSimple_direct_UInt32_ Preparing RedisHash_complex_key_hashed_UInt32_ Prepared RedisHash_complex_key_hashed_UInt32_ Preparing RedisHash_complex_key_cache_UInt32_ Prepared RedisHash_complex_key_cache_UInt32_ Preparing RedisHash_complex_key_direct_UInt32_ Prepared RedisHash_complex_key_direct_UInt32_ Preparing RedisSimple_flat_UInt64_ Prepared RedisSimple_flat_UInt64_ Preparing RedisSimple_hashed_UInt64_ Prepared RedisSimple_hashed_UInt64_ Preparing RedisSimple_cache_UInt64_ Prepared RedisSimple_cache_UInt64_ Preparing RedisSimple_direct_UInt64_ Prepared RedisSimple_direct_UInt64_ Preparing RedisHash_complex_key_hashed_UInt64_ Prepared RedisHash_complex_key_hashed_UInt64_ Preparing RedisHash_complex_key_cache_UInt64_ Prepared RedisHash_complex_key_cache_UInt64_ Preparing RedisHash_complex_key_direct_UInt64_ Prepared RedisHash_complex_key_direct_UInt64_ Preparing RedisSimple_flat_Int8_ Prepared RedisSimple_flat_Int8_ Preparing RedisSimple_hashed_Int8_ Prepared RedisSimple_hashed_Int8_ Preparing RedisSimple_cache_Int8_ Prepared RedisSimple_cache_Int8_ Preparing RedisSimple_direct_Int8_ Prepared RedisSimple_direct_Int8_ Preparing RedisHash_complex_key_hashed_Int8_ Prepared RedisHash_complex_key_hashed_Int8_ Preparing RedisHash_complex_key_cache_Int8_ Prepared RedisHash_complex_key_cache_Int8_ Preparing RedisHash_complex_key_direct_Int8_ Prepared RedisHash_complex_key_direct_Int8_ Preparing RedisSimple_flat_Int16_ Prepared RedisSimple_flat_Int16_ Preparing RedisSimple_hashed_Int16_ Prepared RedisSimple_hashed_Int16_ Preparing RedisSimple_cache_Int16_ Prepared RedisSimple_cache_Int16_ Preparing RedisSimple_direct_Int16_ Prepared RedisSimple_direct_Int16_ Preparing RedisHash_complex_key_hashed_Int16_ Prepared RedisHash_complex_key_hashed_Int16_ Preparing RedisHash_complex_key_cache_Int16_ Prepared RedisHash_complex_key_cache_Int16_ Preparing RedisHash_complex_key_direct_Int16_ Prepared RedisHash_complex_key_direct_Int16_ Preparing RedisSimple_flat_Int32_ Prepared RedisSimple_flat_Int32_ Preparing RedisSimple_hashed_Int32_ Prepared RedisSimple_hashed_Int32_ Preparing RedisSimple_cache_Int32_ Prepared RedisSimple_cache_Int32_ Preparing RedisSimple_direct_Int32_ Prepared RedisSimple_direct_Int32_ Preparing RedisHash_complex_key_hashed_Int32_ Prepared RedisHash_complex_key_hashed_Int32_ Preparing RedisHash_complex_key_cache_Int32_ Prepared RedisHash_complex_key_cache_Int32_ Preparing RedisHash_complex_key_direct_Int32_ Prepared RedisHash_complex_key_direct_Int32_ Preparing RedisSimple_flat_Int64_ Prepared RedisSimple_flat_Int64_ Preparing RedisSimple_hashed_Int64_ Prepared RedisSimple_hashed_Int64_ Preparing RedisSimple_cache_Int64_ Prepared RedisSimple_cache_Int64_ Preparing RedisSimple_direct_Int64_ Prepared RedisSimple_direct_Int64_ Preparing RedisHash_complex_key_hashed_Int64_ Prepared RedisHash_complex_key_hashed_Int64_ Preparing RedisHash_complex_key_cache_Int64_ Prepared RedisHash_complex_key_cache_Int64_ Preparing RedisHash_complex_key_direct_Int64_ Prepared RedisHash_complex_key_direct_Int64_ Preparing RedisSimple_flat_UUID_ Prepared RedisSimple_flat_UUID_ Preparing RedisSimple_hashed_UUID_ Prepared RedisSimple_hashed_UUID_ Preparing RedisSimple_cache_UUID_ Prepared RedisSimple_cache_UUID_ Preparing RedisSimple_direct_UUID_ Prepared RedisSimple_direct_UUID_ Preparing RedisHash_complex_key_hashed_UUID_ Prepared RedisHash_complex_key_hashed_UUID_ Preparing RedisHash_complex_key_cache_UUID_ Prepared RedisHash_complex_key_cache_UUID_ Preparing RedisHash_complex_key_direct_UUID_ Prepared RedisHash_complex_key_direct_UUID_ Preparing RedisSimple_flat_Date_ Prepared RedisSimple_flat_Date_ Preparing RedisSimple_hashed_Date_ Prepared RedisSimple_hashed_Date_ Preparing RedisSimple_cache_Date_ Prepared RedisSimple_cache_Date_ Preparing RedisSimple_direct_Date_ Prepared RedisSimple_direct_Date_ Preparing RedisHash_complex_key_hashed_Date_ Prepared RedisHash_complex_key_hashed_Date_ Preparing RedisHash_complex_key_cache_Date_ Prepared RedisHash_complex_key_cache_Date_ Preparing RedisHash_complex_key_direct_Date_ Prepared RedisHash_complex_key_direct_Date_ Preparing RedisSimple_flat_DateTime_ Prepared RedisSimple_flat_DateTime_ Preparing RedisSimple_hashed_DateTime_ Prepared RedisSimple_hashed_DateTime_ Preparing RedisSimple_cache_DateTime_ Prepared RedisSimple_cache_DateTime_ Preparing RedisSimple_direct_DateTime_ Prepared RedisSimple_direct_DateTime_ Preparing RedisHash_complex_key_hashed_DateTime_ Prepared RedisHash_complex_key_hashed_DateTime_ Preparing RedisHash_complex_key_cache_DateTime_ Prepared RedisHash_complex_key_cache_DateTime_ Preparing RedisHash_complex_key_direct_DateTime_ Prepared RedisHash_complex_key_direct_DateTime_ Preparing RedisSimple_flat_String_ Prepared RedisSimple_flat_String_ Preparing RedisSimple_hashed_String_ Prepared RedisSimple_hashed_String_ Preparing RedisSimple_cache_String_ Prepared RedisSimple_cache_String_ Preparing RedisSimple_direct_String_ Prepared RedisSimple_direct_String_ Preparing RedisHash_complex_key_hashed_String_ Prepared RedisHash_complex_key_hashed_String_ Preparing RedisHash_complex_key_cache_String_ Prepared RedisHash_complex_key_cache_String_ Preparing RedisHash_complex_key_direct_String_ Prepared RedisHash_complex_key_direct_String_ Preparing RedisSimple_flat_Float32_ Prepared RedisSimple_flat_Float32_ Preparing RedisSimple_hashed_Float32_ Prepared RedisSimple_hashed_Float32_ Preparing RedisSimple_cache_Float32_ Prepared RedisSimple_cache_Float32_ Preparing RedisSimple_direct_Float32_ Prepared RedisSimple_direct_Float32_ Preparing RedisHash_complex_key_hashed_Float32_ Prepared RedisHash_complex_key_hashed_Float32_ Preparing RedisHash_complex_key_cache_Float32_ Prepared RedisHash_complex_key_cache_Float32_ Preparing RedisHash_complex_key_direct_Float32_ Prepared RedisHash_complex_key_direct_Float32_ Preparing RedisSimple_flat_Float64_ Prepared RedisSimple_flat_Float64_ Preparing RedisSimple_hashed_Float64_ Prepared RedisSimple_hashed_Float64_ Preparing RedisSimple_cache_Float64_ Prepared RedisSimple_cache_Float64_ Preparing RedisSimple_direct_Float64_ Prepared RedisSimple_direct_Float64_ Preparing RedisHash_complex_key_hashed_Float64_ Prepared RedisHash_complex_key_hashed_Float64_ Preparing RedisHash_complex_key_cache_Float64_ Prepared RedisHash_complex_key_cache_Float64_ Preparing RedisHash_complex_key_direct_Float64_ Prepared RedisHash_complex_key_direct_Float64_ Run test with id: 10 Executing query system reload dictionaries on node Executing query SELECT engine_full FROM system.tables WHERE name = 'test_experimental' on instance Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(4), 4) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_database_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table_1 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_1 VALUES (0, 'test_database.test_table_1'); CREATE TABLE test_database.test_table_2 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_2 VALUES (0, 'test_database.test_table_2'); CREATE TABLE test_database.test_table_3 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_3 VALUES (0, 'test_database.test_table_3'); BACKUP DATABASE test_database TO Disk('backup_disk_object_storage_local_plain', 'test_database_backup'); CREATE DATABASE test_database_backup ENGINE = Backup('test_database', Disk('backup_disk_object_storage_local_plain', 'test_database_backup')); on instance Executing query CREATE TABLE test_experimental_new (uid String, version UInt32, is_deleted UInt8) ENGINE = ReplacingMergeTree(version, is_deleted) ORDER by (uid) SETTINGS allow_experimental_replacing_merge_with_cleanup=1; on instance Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select value from system.settings where name = 'max_query_size' on node7 Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(33), 33) on node Executing query RESTORE TABLE tbl FROM Disk('backups', '2') on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(34), 34) on node Executing query rename table system.text_log to system.text_log_8_test on node2 Executing query select value from system.settings where name = 'max_threads' on node7 Executing query CREATE TABLE test_experimental_new (uid String, version UInt32, is_deleted UInt8) ENGINE = ReplacingMergeTree(version, is_deleted) ORDER by (uid); on instance Executing query SELECT * FROM tbl on node1 Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query rename table system.query_log to system.query_log_8_test on node2 Executing query system reload dictionary RedisSimple_flat_DateTime_ on node Executing query select value from system.settings where name = 'max_query_size' on node8 Executing query SELECT * FROM test.tbl3 ORDER BY p on node Executing query ALTER TABLE test_experimental_new MODIFY setting allow_experimental_replacing_merge_with_cleanup=1 on instance Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] Executing query system flush logs on node2 Executing query select dictGet('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") [gw7] PASSED test_config_substitutions/test.py::test_config run container_id:roottestconfigsubstitutions-gw7-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_a1ccb17e-c509-432d-82de-2da8f32f0a82.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node7-1 bash -c test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_a1ccb17e-c509-432d-82de-2da8f32f0a82.xml] test_config_substitutions/test.py::test_config_from_env_overrides run container_id:roottestconfigsubstitutions-gw7-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n\n \n \n \n 100\n \n \n \n \n \n default\n default\n \n\n \n \n \n\n\' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node7-1 bash -c echo ' 100 default default ' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] copy from /var/lib/clickhouse/shadow/*/data/test/tbl to /var/lib/clickhouse/data/test/tbl3/detached run container_id:roottestbackuprestore-gw6-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl3/detached'] Command:[docker exec roottestbackuprestore-gw6-node-1 bash -c cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl3/detached] Executing query DROP TABLE IF EXISTS test_experimental_new on instance Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query ALTER TABLE test.tbl3 REPLACE PARTITION 197002 FROM test.tbl on node Executing query SYSTEM RELOAD CONFIG on node7 Executing query select dictGetDateTime('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(1)) on node Executing query select dictGetInt16('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query SELECT sleep(2) on node Executing query select dictHas('RedisSimple_flat_DateTime_', toUInt64(1)) on node Executing query SYSTEM RELOAD CONFIG on instance run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Stdout: PID TTY TIME CMD Stdout: 796 ? 00:00:03 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] run container_id:roottestconfigsubstitutions-gw7-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/000-users_with_env_subst_a1ccb17e-c509-432d-82de-2da8f32f0a82.xml || mv /tmp/000-users_with_env_subst_a1ccb17e-c509-432d-82de-2da8f32f0a82.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node7-1 bash -c test ! -f /tmp/000-users_with_env_subst_a1ccb17e-c509-432d-82de-2da8f32f0a82.xml || mv /tmp/000-users_with_env_subst_a1ccb17e-c509-432d-82de-2da8f32f0a82.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestconfigsubstitutions-gw7-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_49b44a77-3c5a-4de5-98e1-962d955f8653.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node7-1 bash -c test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_49b44a77-3c5a-4de5-98e1-962d955f8653.xml] Stdout:796 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node run container_id:roottestconfigsubstitutions-gw7-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n\n \n \n \n 1\n \n \n \n \n \n default\n default\n \n\n \n \n \n\n\' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node7-1 bash -c echo ' 1 default default ' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query SYSTEM RELOAD CONFIG on node7 Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query select dictGetDateTimeOrDefault('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node run container_id:roottestconfigsubstitutions-gw7-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/000-users_with_env_subst_49b44a77-3c5a-4de5-98e1-962d955f8653.xml || mv /tmp/000-users_with_env_subst_49b44a77-3c5a-4de5-98e1-962d955f8653.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node7-1 bash -c test ! -f /tmp/000-users_with_env_subst_49b44a77-3c5a-4de5-98e1-962d955f8653.xml || mv /tmp/000-users_with_env_subst_49b44a77-3c5a-4de5-98e1-962d955f8653.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] Executing query DROP TABLE IF EXISTS test_experimental on instance Executing query rename table system.text_log to system.text_log_9_test on node2 [gw7] PASSED test_config_substitutions/test.py::test_config_from_env_overrides Executing query SYSTEM RELOAD CONFIG on node7 test_config_substitutions/test.py::test_config_merge_from_env_overrides Executing query select dictGet('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(2)) on node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance [gw8] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query rename table system.query_log to system.query_log_9_test on node2 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetDateTime('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(2)) on node Executing query DROP TABLE IF EXISTS test_experimental on instance Executing query system flush logs on node2 Executing query SELECT value FROM system.server_settings WHERE name='max_thread_pool_size' on node7 Executing query select dictGetInt32('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Stdout:796 Executing query select dictHas('RedisSimple_flat_DateTime_', toUInt64(2)) on node Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] run container_id:roottestconfigsubstitutions-gw7-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/config.d/010-server_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/config.d/010-server_with_env_subst.xml /tmp/010-server_with_env_subst_392d66e9-6a89-41c9-8c50-7acc09ec6837.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node7-1 bash -c test ! -f /etc/clickhouse-server/config.d/010-server_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/config.d/010-server_with_env_subst.xml /tmp/010-server_with_env_subst_392d66e9-6a89-41c9-8c50-7acc09ec6837.xml] Executing query SYSTEM RELOAD CONFIG on instance Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") run container_id:roottestconfigsubstitutions-gw7-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n\n 9000\n\n\' > /etc/clickhouse-server/config.d/010-server_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node7-1 bash -c echo ' 9000 ' > /etc/clickhouse-server/config.d/010-server_with_env_subst.xml] Executing query SYSTEM RELOAD CONFIG on node7 Executing query select dictGetOrDefault('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query INSERT INTO tbl VALUES (1) on node1 run container_id:roottestconfigsubstitutions-gw7-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/010-server_with_env_subst_392d66e9-6a89-41c9-8c50-7acc09ec6837.xml || mv /tmp/010-server_with_env_subst_392d66e9-6a89-41c9-8c50-7acc09ec6837.xml /etc/clickhouse-server/config.d/010-server_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node7-1 bash -c test ! -f /tmp/010-server_with_env_subst_392d66e9-6a89-41c9-8c50-7acc09ec6837.xml || mv /tmp/010-server_with_env_subst_392d66e9-6a89-41c9-8c50-7acc09ec6837.xml /etc/clickhouse-server/config.d/010-server_with_env_subst.xml] Executing query select dictGetDateTimeOrDefault('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node [gw7] PASSED test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestconfigsubstitutions-gw7-node3-1 bash -c ps -C clickhouse] Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query create table b (a Int64) ENGINE=MergeTree() order by a; on instance Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:03 clickhouse run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestconfigsubstitutions-gw7-node3-1 bash -c pkill clickhouse] run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw7-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '3') ASYNC on node1 Stdout:8 Executing query SELECT * FROM test.tbl3 ORDER BY p on node Executing query system reload dictionary RedisSimple_hashed_DateTime_ on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query rename table system.text_log to system.text_log_10_test on node2 Executing query SYSTEM RELOAD CONFIG on instance run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '3') ASYNC on node2 Stdout:796 Executing query select dictGet('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(1)) on node Executing query rename table system.query_log to system.query_log_10_test on node2 Executing query ALTER TABLE test.tbl3 UPDATE k=10 WHERE 1 on node Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query select dictGetInt64('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query SELECT status FROM system.backups WHERE id IN ['717d4b5a-0393-402d-8d1f-4837f7acdc00', 'eba50924-f04b-405b-bd1a-a6ca01ab276a'] AND status == 'CREATING_BACKUP' on node1 Executing query select dictGetDateTime('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(1)) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node2-1 bash -c ps -C clickhouse] Executing query SELECT sleep(2) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:12 clickhouse run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node2-1 bash -c pkill clickhouse] Executing query DROP TABLE IF EXISTS b on instance run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") [gw4] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] Executing query SELECT hostName() FROM clusterAllReplicas('two_shards_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=0 on node1 Executing query select dictHas('RedisSimple_hashed_DateTime_', toUInt64(1)) on node [gw8] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user Executing query DROP USER IF EXISTS user_experimental on instance Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw7-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetOrDefault('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT status FROM system.backups WHERE id IN ['717d4b5a-0393-402d-8d1f-4837f7acdc00', 'eba50924-f04b-405b-bd1a-a6ca01ab276a'] AND status == 'CREATING_BACKUP' on node1 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Stdout:796 Executing query select dictGetDateTimeOrDefault('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SYSTEM RELOAD CONFIG on instance Executing query SELECT status FROM system.backups WHERE id IN ['717d4b5a-0393-402d-8d1f-4837f7acdc00', 'eba50924-f04b-405b-bd1a-a6ca01ab276a'] AND status == 'CREATING_BACKUP' on node2 Executing query select dictGet('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(2)) on node Executing query select dictGetUUID('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Stdout:8 Executing query SELECT count() FROM system.backups WHERE id IN ['717d4b5a-0393-402d-8d1f-4837f7acdc00', 'eba50924-f04b-405b-bd1a-a6ca01ab276a'] AND status == 'BACKUP_CREATED' on node1 Executing query select dictGetDateTime('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query CREATE USER user_experimental IDENTIFIED WITH no_password SETTINGS allow_experimental_time_series_table = 1 on instance run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw7-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query SELECT count() FROM system.backups WHERE id IN ['717d4b5a-0393-402d-8d1f-4837f7acdc00', 'eba50924-f04b-405b-bd1a-a6ca01ab276a'] AND status == 'BACKUP_CREATED' on node2 Executing query select dictHas('RedisSimple_hashed_DateTime_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SYSTEM RELOAD CONFIG on instance run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT count() FROM system.backups WHERE id IN ['717d4b5a-0393-402d-8d1f-4837f7acdc00', 'eba50924-f04b-405b-bd1a-a6ca01ab276a'] AND status == 'BACKUP_FAILED' on node1 Executing query select dictGetOrDefault('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/d809932b30fd2841270b84cf696055981f9f5667df058f91ebaa0d12dacc1808/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/d809932b30fd2841270b84cf696055981f9f5667df058f91ebaa0d12dacc1808/json HTTP/1.1" 200 586 Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query SELECT count() FROM system.backups WHERE id IN ['717d4b5a-0393-402d-8d1f-4837f7acdc00', 'eba50924-f04b-405b-bd1a-a6ca01ab276a'] AND status == 'BACKUP_FAILED' on node2 Executing query select dictGetDateTimeOrDefault('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT * FROM test.tbl3 ORDER BY p on node Stdout:8 Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query system reload dictionary RedisSimple_cache_DateTime_ on node Executing query CREATE USER user_experimental IDENTIFIED WITH no_password SETTINGS allow_experimental_time_series_table = 1 on instance run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw7-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetDate('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node Stdout:8 Executing query DROP TABLE IF EXISTS test.tbl3 on node Executing query SELECT value FROM system.settings WHERE name = 'allow_experimental_time_series_table' on instance Executing query select dictGet('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(1)) on node Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query RESTORE TABLE tbl FROM Disk('backups', '3') on node1 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node [gw6] PASSED test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore Executing query CREATE TABLE test.tbl1 AS test.tbl on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT value FROM system.settings WHERE name = 'allow_experimental_time_series_table' on instance Executing query select dictGetDateTime('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(1)) on node Stdout:1617 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1617 Executing query select 20 on instance Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node copy from /var/lib/clickhouse/shadow/*/data/test/tbl to /var/lib/clickhouse/data/test/tbl1/detached run container_id:roottestbackuprestore-gw6-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl1/detached'] Command:[docker exec roottestbackuprestore-gw6-node-1 bash -c cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl1/detached] Executing query ALTER TABLE test.tbl1 ATTACH PARTITION 197001 on node Executing query SELECT * FROM tbl on node1 Executing query select dictHas('RedisSimple_cache_DateTime_', toUInt64(1)) on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query SYSTEM RELOAD CONFIG on instance run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query ALTER TABLE test.tbl1 ATTACH PARTITION 197002 on node Stdout:8 Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] Executing query select dictGetOrDefault('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query SELECT sleep(2) on node run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw7-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw7-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestconfigsubstitutions-gw7-node3-1/exec HTTP/1.1" 201 74 Executing query select dictGetDateTimeOrDefault('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query select 20 on instance http://localhost:None "POST /v1.46/exec/201d12de23e25a160b322798105cc0318a7852f33556100fffd47037ee1b7a4e/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/201d12de23e25a160b322798105cc0318a7852f33556100fffd47037ee1b7a4e/json HTTP/1.1" 200 586 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestallowfeaturetier-gw8-instance-1 bash -c ps -C clickhouse] Executing query select dictGetDateTime('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGet('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(2)) on node Stdout: PID TTY TIME CMD Stdout: 777 ? 00:00:04 clickhouse run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestallowfeaturetier-gw8-instance-1 bash -c pkill clickhouse] run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:777 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetDateTime('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(2)) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select 20 on instance Stdout:8 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictHas('RedisSimple_cache_DateTime_', toUInt64(2)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw7-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Stdout:797 Clickhouse process running. run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw7-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:797 Executing query select 20 on node3 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Executing query select dictGetDateTimeOrDefault('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Stdout:777 Mysql 8 Started ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/.env --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw5 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/.env --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw5 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml up -d --no-recreate] Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query select dictGetString('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node Executing query system reload dictionary RedisSimple_direct_DateTime_ on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(1)) on node Executing query CREATE TABLE mydb.test (`name` String, `value` UInt32) ENGINE = ReplicatedMergeTree ORDER BY value on node1 Executing query DROP DATABASE test_database_backup on instance Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw9-node2-1/exec HTTP/1.1" 201 74 Executing query select 20 on node3 http://localhost:None "POST /v1.46/exec/7d5e2afe036f2a5b579a02cc28675bbdd9c30ef74b326c457b826d8c374f0b45/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/7d5e2afe036f2a5b579a02cc28675bbdd9c30ef74b326c457b826d8c374f0b45/json HTTP/1.1" 200 586 Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1 Running Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1 Creating Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1 Created Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1 Starting Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.10.3... http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None Executing query SELECT * FROM test.tbl1 ORDER BY p on node Executing query DROP DATABASE test_database on instance Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query select dictGetDateTime('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None Executing query INSERT INTO mydb.test VALUES ('abc', 1), ('def', 2) on node1 Executing query ALTER TABLE test.tbl1 UPDATE k=10 WHERE 1 on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query select dictHas('RedisSimple_direct_DateTime_', toUInt64(1)) on node http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Stdout:777 http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_database_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table_1 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_1 VALUES (0, 'test_database.test_table_1'); CREATE TABLE test_database.test_table_2 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_2 VALUES (0, 'test_database.test_table_2'); CREATE TABLE test_database.test_table_3 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_3 VALUES (0, 'test_database.test_table_3'); BACKUP DATABASE test_database TO Disk('backup_disk_s3_plain', 'test_database_backup'); CREATE DATABASE test_database_backup ENGINE = Backup('test_database', Disk('backup_disk_s3_plain', 'test_database_backup')); on instance http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None Executing query select 20 on node3 Executing query SELECT sleep(2) on node Executing query INSERT INTO mydb.test VALUES ('ghi', 3) on node1 Executing query select dictGetOrDefault('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None [gw4] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] test_cluster_all_replicas/test.py::test_global_in Executing query DROP TABLE IF EXISTS u; on node1 http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None Executing query SYSTEM RELOAD CONFIG on node3 run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None Executing query BACKUP DATABASE mydb TO Disk('backups', '4') on node1 Executing query select dictGetDateTimeOrDefault('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetFloat32('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query CREATE TABLE u(uid Int16) ENGINE=Memory as select 0 on node1 Stdout:788 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None Stdout:788 Executing query select 20 on node2 http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT value FROM system.merge_tree_settings WHERE name='min_bytes_for_wide_part' on node3 http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None Executing query select dictGet('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(2)) on node Executing query SELECT hostName(), * FROM clusterAllReplicas("one_shard_two_nodes", system.one) where dummy GLOBAL IN u ORDER BY ALL on node1 http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None Executing query DROP DATABASE mydb SYNC on node1 Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query SELECT value FROM system.merge_tree_settings WHERE name='min_rows_for_wide_part' on node3 Executing query select dictGetDateTime('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(2)) on node http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None [gw4] PASSED test_cluster_all_replicas/test.py::test_global_in test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] Executing query SELECT hostName() FROM clusterAllReplicas('one_shard_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=1 on node1 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query RESTORE DATABASE mydb FROM Disk('backups', '4') on node1 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestallowfeaturetier-gw8-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/ef648a9e2f38769595712432d7270b04a2656bc0f25ca5bbef6eca676776d8a2/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/ef648a9e2f38769595712432d7270b04a2656bc0f25ca5bbef6eca676776d8a2/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/containers/2b0a69a2820adaf10ba07a77e8cb33021818469a11569175a3438277233f4c4f/json HTTP/1.1" 200 None Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node ClickHouse node started pymysql connect root, clickhouse, 172.16.10.2, 3306 Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Executing query select 20 on node2 Executing query SELECT value FROM system.merge_tree_settings WHERE name='merge_max_block_size' on node3 Executing query select dictHas('RedisSimple_direct_DateTime_', toUInt64(2)) on node pymysql connect root, clickhouse, 172.16.10.2, 3306 pymysql connect root, clickhouse, 172.16.10.2, 3306 pymysql connect root, clickhouse, 172.16.10.2, 3306 pymysql connect root, clickhouse, 172.16.10.2, 3306 pymysql connect root, clickhouse, 172.16.10.2, 3306 pymysql connect root, clickhouse, 172.16.10.2, 3306 Executing query select dictGetOrDefault('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance pymysql connect root, clickhouse, 172.16.10.2, 3306 Executing query SELECT value FROM system.server_settings WHERE name='background_pool_size' on node3 Executing query system reload dictionaries on node Executing query RESTORE DATABASE mydb FROM Disk('backups', '4') SETTINGS structure_only=true on node1 Executing query select dictGetFloat64('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetDateTimeOrDefault('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw7-zoo1-1/json HTTP/1.1" 200 None Executing query select dictGet('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node get_kazoo_client: zoo1, ip:172.16.6.2, port:2181, use_ssl:False run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): Create(path='/background_pool_size', data=b'72', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=1): '/background_pool_size' run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/config.d/config_zk_include_test.xml || mv --no-clobber /etc/clickhouse-server/config.d/config_zk_include_test.xml /tmp/config_zk_include_test_30a30b19-3069-48f9-923a-2b936bcf9797.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node3-1 bash -c test ! -f /etc/clickhouse-server/config.d/config_zk_include_test.xml || mv --no-clobber /etc/clickhouse-server/config.d/config_zk_include_test.xml /tmp/config_zk_include_test_30a30b19-3069-48f9-923a-2b936bcf9797.xml] Stdout: PID TTY TIME CMD Stdout: 1617 ? 00:00:03 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] Executing query select 20 on node2 run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n \n \n 44\n \n \n 1\n 1111\n \n\n \n \n \' > /etc/clickhouse-server/config.d/config_zk_include_test.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node3-1 bash -c echo ' 44 1 1111 ' > /etc/clickhouse-server/config.d/config_zk_include_test.xml] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SYSTEM RELOAD CONFIG on node3 Stdout:1617 Executing query SYSTEM SYNC REPLICA mydb.test on node1 Executing query system reload dictionary RedisHash_complex_key_hashed_DateTime_ on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUInt8('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1589 Clickhouse process running. run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT * FROM mydb.test ORDER BY name on node1 Stdout:1589 Executing query select 20 on instance Executing query SELECT * FROM test.tbl1 ORDER BY p on node Executing query select dictGet('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT value FROM system.server_settings WHERE name='background_pool_size' on node3 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select count() from system.text_log_5_test on node2 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP DATABASE mydb SYNC on node1 Executing query DROP TABLE IF EXISTS test.tbl1 on node run container_id:roottestconfigsubstitutions-gw7-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/config_zk_include_test_30a30b19-3069-48f9-923a-2b936bcf9797.xml || mv /tmp/config_zk_include_test_30a30b19-3069-48f9-923a-2b936bcf9797.xml /etc/clickhouse-server/config.d/config_zk_include_test.xml'] Command:[docker exec roottestconfigsubstitutions-gw7-node3-1 bash -c test ! -f /tmp/config_zk_include_test_30a30b19-3069-48f9-923a-2b936bcf9797.xml || mv /tmp/config_zk_include_test_30a30b19-3069-48f9-923a-2b936bcf9797.xml /etc/clickhouse-server/config.d/config_zk_include_test.xml] Executing query select dictGetDateTime('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Sending request(xid=2): Delete(path='/background_pool_size', version=-1) Received response(xid=2): True Executing query select 1 on node4 [gw7] PASSED test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config Executing query select count() from system.query_log_1_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/.env --project-name roottestbackuprestore-gw6 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/node/docker-compose.yml stop --timeout 20] [gw6] PASSED test_backup_restore/test.py::test_restore Executing query RESTORE DATABASE mydb FROM Disk('backups', '4') SETTINGS allow_non_empty_tables=true on node1 Executing query select dictHas('RedisHash_complex_key_hashed_DateTime_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stdout:1617 Executing query select 1 on node4 Executing query select 20 on instance Executing query select count() from system.text_log_8_test on node2 Executing query select dictGetUInt8OrDefault('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetUInt8('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select 1 on node4 Executing query SYSTEM SYNC REPLICA mydb.test on node1 Executing query select count() from system.query_log_10_test on node2 Executing query select dictGet('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SELECT * FROM mydb.test ORDER BY name on node1 Executing query select 1 on node3 Executing query select count() from system.text_log_2_test on node2 Executing query select dictGetUInt16('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select 20 on instance Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica Executing query select 1 on node3 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.query_log_2_test on node2 Executing query select dictGetDateTime('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Stdout:1617 Executing query SELECT value FROM system.settings WHERE name = 'allow_experimental_time_series_table' on instance Executing query select 1 on node3 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select count() from system.asynchronous_loader where job ilike '%_log_%_test' and execution_pool = 'BackgroundLoad' on node2 Executing query select dictHas('RedisHash_complex_key_hashed_DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SELECT value FROM system.settings WHERE name = 'allow_experimental_time_series_table' on instance Executing query select 1 on node8 Executing query select dictGetUInt16OrDefault('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query system flush logs on node2 Executing query select dictGetUInt16('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select 1 on node8 Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=1 on instance Executing query select dictGet('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select 1 on node8 Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=0 on instance Executing query select dictGetUInt32('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query system reload dictionary RedisHash_complex_key_cache_DateTime_ on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1617 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/.env --project-name roottestconfigsubstitutions-gw7 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node8/docker-compose.yml stop --timeout 20] [gw7] PASSED test_config_substitutions/test.py::test_include_config run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query SYSTEM RELOAD CONFIG on instance Executing query select dictGet('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query rename table system.text_log to system.text_log_11_test on node2 Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query select dictGetDateTime('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes Executing query CREATE TABLE tbl (`x` UInt8, `y` String) ENGINE = MergeTree ORDER BY x on node1 Executing query select dictGetUInt32OrDefault('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query rename table system.query_log to system.query_log_11_test on node2 Executing query CREATE TABLE tbl (`w` Int64) ENGINE = MergeTree ORDER BY w on node2 Executing query select dictGetUInt32('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP USER IF EXISTS user_experimental on instance Executing query select dictHas('RedisHash_complex_key_cache_DateTime_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query system flush logs on node2 [gw8] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_user run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/2/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/2/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query select dictGet('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query INSERT INTO tbl VALUES (1, 'Don''t'), (2, 'count'), (3, 'your'), (4, 'chickens') on node1 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stdout:1617 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/allow_experimental_time_series_table>./allow_experimental_time_series_table>1/g' /etc/clickhouse-server/users.d/users.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/allow_experimental_time_series_table>./allow_experimental_time_series_table>1/g' /etc/clickhouse-server/users.d/users.xml] Executing query SYSTEM RELOAD CONFIG on instance Executing query select dictGetUInt64('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query INSERT INTO tbl VALUES (-333), (-222), (-111), (0), (111) on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '5') on node1 Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 [gw4] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] Executing query SELECT hostName() FROM clusterAllReplicas('two_shards_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=1 on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 Stderr: Container roottestbackuprestore-gw6-node-1 Stopping Stderr: Container roottestbackuprestore-gw6-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/.env --project-name roottestbackuprestore-gw6 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw6/node/docker-compose.yml down --volumes] http://localhost:None "POST /v1.46/exec/20b26a560d3a08f955b80cd82dd6ad5ee31a239ef43b443d909a3c4d30bda5fb/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/20b26a560d3a08f955b80cd82dd6ad5ee31a239ef43b443d909a3c4d30bda5fb/json HTTP/1.1" 200 586 Executing query rename table system.text_log to system.text_log_12_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictGetDateTime('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SELECT value FROM system.settings WHERE name = 'allow_experimental_time_series_table' on instance Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query rename table system.query_log to system.query_log_12_test on node2 Executing query select dictGetUInt64OrDefault('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictHas('RedisHash_complex_key_cache_DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt64('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=0 on instance Executing query system flush logs on node2 Executing query select dictGet('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '5') on node2 Executing query select dictGetOrDefault('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stderr: Container roottestbackuprestore-gw6-node-1 Stopping Stderr: Container roottestbackuprestore-gw6-node-1 Stopped Stderr: Container roottestbackuprestore-gw6-node-1 Removing Stderr: Container roottestbackuprestore-gw6-node-1 Removed Stderr: Network roottestbackuprestore-gw6_default Removing Stderr: Network roottestbackuprestore-gw6_default Removed Cleanup called Docker networks for project roottestbackuprestore-gw6 are NETWORK ID NAME DRIVER SCOPE run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/2/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/2/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Docker containers for project roottestbackuprestore-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackuprestore-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestore-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackuprestore-gw6 Trying to prune unused networks... run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/allow_experimental_time_series_table>./allow_experimental_time_series_table>0/g' /etc/clickhouse-server/users.d/users.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/allow_experimental_time_series_table>./allow_experimental_time_series_table>0/g' /etc/clickhouse-server/users.d/users.xml] Trying to prune unused images... Command:[docker image prune -f] Executing query SYSTEM RELOAD CONFIG on instance Executing query select dictGetInt8('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Running tests in /ClickHouse/tests/integration/test_composable_protocols/test.py test_composable_protocols/test.py::test_connections Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Cluster start called. is_up=False Docker networks for project roottestcomposableprotocols-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcomposableprotocols-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Docker volumes for project roottestcomposableprotocols-gw6 are DRIVER VOLUME NAME Cleanup called Stdout:2475 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Docker networks for project roottestcomposableprotocols-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcomposableprotocols-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcomposableprotocols-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcomposableprotocols-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestcomposableprotocols-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:2475 Executing query select 20 on instance Executing query SELECT * FROM tbl on node1 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: server Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_composable_protocols/configs/server.crt', '/ClickHouse/tests/integration/test_composable_protocols/configs/server.key'] to /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/server/configs/config.d Setup database dir /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/server/database Setup logs dir /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/server/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/.env --project-name roottestcomposableprotocols-gw6 --file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/server/docker-compose.yml pull] Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query system reload dictionary RedisHash_complex_key_direct_DateTime_ on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query SELECT * FROM tbl on node2 Executing query select dictGet('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query rename table system.text_log to system.text_log_13_test on node2 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Command:[docker compose --env-file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env --project-name roottestallowfeaturetier-gw8 --file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/docker-compose.yml stop --timeout 20] [gw8] PASSED test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile Executing query select dictGetDateTime('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query rename table system.query_log to system.query_log_13_test on node2 Executing query select dictGetInt8OrDefault('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetInt8('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select 20 on instance Executing query system flush logs on node2 Stderr: Container roottestallowfeaturetier-gw8-instance-1 Stopping Stderr: Container roottestallowfeaturetier-gw8-instance-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env --project-name roottestallowfeaturetier-gw8 --file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/docker-compose.yml down --volumes] Executing query select dictHas('RedisHash_complex_key_direct_DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt16('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on instance Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Stderr: Container roottestallowfeaturetier-gw8-instance-1 Stopping Stderr: Container roottestallowfeaturetier-gw8-instance-1 Stopped Stderr: Container roottestallowfeaturetier-gw8-instance-1 Removing Stderr: Container roottestallowfeaturetier-gw8-instance-1 Removed Stderr: Network roottestallowfeaturetier-gw8_default Removing Stderr: Network roottestallowfeaturetier-gw8_default Removed Cleanup called Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Docker networks for project roottestallowfeaturetier-gw8 are NETWORK ID NAME DRIVER SCOPE Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Docker containers for project roottestallowfeaturetier-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Docker volumes for project roottestallowfeaturetier-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestallowfeaturetier-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestallowfeaturetier-gw8 Trying to prune unused networks... Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query select dictGet('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Running tests in /ClickHouse/tests/integration/test_config_decryption/test.py test_config_decryption/test.py::test_successful_decryption_xml Cluster start called. is_up=False Docker networks for project roottestconfigdecryption-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfigdecryption-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query rename table system.text_log to system.text_log_14_test on node2 Docker volumes for project roottestconfigdecryption-gw8 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestconfigdecryption-gw8 are NETWORK ID NAME DRIVER SCOPE Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Docker containers for project roottestconfigdecryption-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigdecryption-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigdecryption-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Unstopped containers: {} No running containers for project: roottestconfigdecryption-gw8 Trying to prune unused networks... Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Trying to prune unused images... Command:[docker image prune -f] Executing query select dictGetDateTime('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Executing query rename table system.query_log to system.query_log_14_test on node2 test_backup_restore_on_cluster/test.py::test_empty_replicated_table Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_decryption/configs/config.xml'] to /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/database Setup logs dir /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_decryption/configs/config.yaml'] to /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/database Setup logs dir /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env --project-name roottestconfigdecryption-gw8 --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/docker-compose.yml pull] Executing query select dictGetInt16OrDefault('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetInt16('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisHash_complex_key_direct_DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query system flush logs on node2 Executing query select dictGet('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '6') SETTINGS replica_num=1 on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query DROP DATABASE test_database_backup on instance Executing query select dictGetInt32('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query DROP DATABASE test_database on instance Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query system reload dictionaries on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/.backup HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/checksums.txt HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/columns.txt HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/count.txt HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/data.bin HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/data.cmrk3 HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/default_compression_codec.txt HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/metadata_version.txt HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/primary.cidx HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/serialization.json HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_2/all_1_1_0/checksums.txt HTTP/1.1" 204 0 Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_2/all_1_1_0/data.bin HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_3/all_1_1_0/checksums.txt HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_3/all_1_1_0/data.bin HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/metadata/test_database.sql HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/metadata/test_database/test_table_1.sql HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/metadata/test_database/test_table_2.sql HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/metadata/test_database/test_table_3.sql HTTP/1.1" 204 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node [gw2] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] Run test with id: 9 Executing query system reload dictionaries on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query rename table system.text_log to system.text_log_15_test on node2 Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '6') on node1 http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_database_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table_1 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_1 VALUES (0, 'test_database.test_table_1'); CREATE TABLE test_database.test_table_2 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_2 VALUES (0, 'test_database.test_table_2'); CREATE TABLE test_database.test_table_3 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_3 VALUES (0, 'test_database.test_table_3'); BACKUP DATABASE test_database TO File('test_database_backup_file'); CREATE DATABASE test_database_backup ENGINE = Backup('test_database', File('test_database_backup_file')); on instance Executing query select dictGetInt32OrDefault('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGetInt32('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query system reload dictionary RedisSimple_flat_Date_ on node Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query rename table system.query_log to system.query_log_15_test on node2 Executing query select dictGet('RedisSimple_flat_Date_', 'Date_', toUInt64(1)) on node Executing query select dictGet('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Executing query system flush logs on node2 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictGetDate('RedisSimple_flat_Date_', 'Date_', toUInt64(1)) on node Executing query select dictGetInt64('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query SELECT * FROM tbl on node1 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query select dictHas('RedisSimple_flat_Date_', toUInt64(1)) on node Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query SELECT * FROM tbl on node2 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: 2475 ? 00:00:04 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Connection dropped: outstanding heartbeat ping not received Transition to CONNECTING Zookeeper connection lost Executing query select dictGetInt64OrDefault('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Stdout:2475 Executing query select dictGetDateOrDefault('RedisSimple_flat_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query rename table system.text_log to system.text_log_16_test on node2 Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Executing query select dictGetInt64('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query rename table system.query_log to system.query_log_16_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisSimple_flat_Date_', 'Date_', toUInt64(2)) on node Executing query select dictGet('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query system flush logs on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query select dictGetUUID('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDate('RedisSimple_flat_Date_', 'Date_', toUInt64(2)) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisSimple_flat_Date_', toUInt64(2)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2475 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetOrDefault('RedisSimple_flat_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetUUID('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetDateOrDefault('RedisSimple_flat_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetUUIDOrDefault('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query rename table system.text_log to system.text_log_17_test on node2 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query system reload dictionary RedisSimple_hashed_Date_ on node Executing query rename table system.query_log to system.query_log_17_test on node2 Executing query select dictGet('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node test_backup_restore_on_cluster/test.py::test_file_deduplication Executing query CREATE TABLE tbl ON CLUSTER 'cluster' ( piPItkORAM Int32 ) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}') ORDER BY tuple() SETTINGS min_bytes_for_wide_part=0 on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2475 Executing query select dictGet('RedisSimple_hashed_Date_', 'Date_', toUInt64(1)) on node Executing query system flush logs on node2 Executing query select dictGetDate('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetDate('RedisSimple_hashed_Date_', 'Date_', toUInt64(1)) on node Executing query CREATE TABLE tbl2 ON CLUSTER 'cluster' ( piPItkORAM Int32 ) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}-2') ORDER BY tuple() SETTINGS min_bytes_for_wide_part=0 on node1 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisSimple_hashed_Date_', toUInt64(1)) on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query INSERT INTO tbl VALUES (3556), (1177), (4004), (4264), (3729), (1438), (2158), (2684), (415), (1917) on node1 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetDate('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_hashed_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetDateOrDefault('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Stdout:2475 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query rename table system.text_log to system.text_log_18_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetDateOrDefault('RedisSimple_hashed_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGet('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query rename table system.query_log to system.query_log_18_test on node2 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl2 on node1 Executing query select dictGet('RedisSimple_hashed_Date_', 'Date_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGetDateTime('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query system flush logs on node2 Executing query select dictGetDate('RedisSimple_hashed_Date_', 'Date_', toUInt64(2)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query BACKUP TABLE tbl, TABLE tbl2 ON CLUSTER 'cluster' TO Disk('backups', '7') on node1 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisSimple_hashed_Date_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Connection dropped: socket connection error: No route to host Connection dropped: socket connection error: No route to host run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2475 Stdout:3274 Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Executing query select dictGetOrDefault('RedisSimple_hashed_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetDateTime('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query SYSTEM FLUSH LOGS ON CLUSTER 'cluster' on node1 Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetDateOrDefault('RedisSimple_hashed_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query rename table system.text_log to system.text_log_19_test on node2 Executing query select dictGet('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query system reload dictionary RedisSimple_cache_Date_ on node Stderr: node1 Skipped - Image is already being pulled by node2 Stderr: node2 Pulling Stderr: node2 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env --project-name roottestconfigdecryption-gw8 --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env --project-name roottestconfigdecryption-gw8 --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/docker-compose.yml up -d --no-recreate] Executing query rename table system.query_log to system.query_log_19_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGet('RedisSimple_cache_Date_', 'Date_', toUInt64(1)) on node Executing query select dictGetString('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(1),toString('world'))) on node Stderr: Container roottestconfigsubstitutions-gw7-node5-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node4-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node6-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node8-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node7-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-zoo3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-zoo1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-zoo2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-zoo3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-zoo2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node7-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node4-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node5-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node8-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node1-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node6-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query system flush logs on node2 Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node4/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node4/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node5/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node5/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node6/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node6/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node7/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node7/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node8/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node8/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/.env --project-name roottestconfigsubstitutions-gw7 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw7/node8/docker-compose.yml down --volumes] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/3886012bddb79bfcf9c8e4f616c576d1a874e7279c70f2b080db9abece0cc5d7/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/3886012bddb79bfcf9c8e4f616c576d1a874e7279c70f2b080db9abece0cc5d7/json HTTP/1.1" 200 586 Executing query select dictGetDate('RedisSimple_cache_Date_', 'Date_', toUInt64(1)) on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*piPItkORAM.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c grep -a "Writing backup for file .*piPItkORAM.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stdout:1 run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*piPItkORAM.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c grep -a "Writing backup for file .*piPItkORAM.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Stdout:0 run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*piPItkORAM.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c grep -a "Writing backup for file .*piPItkORAM.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Stdout:3 run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*piPItkORAM.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c grep -a "Writing backup for file .*piPItkORAM.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Executing query select dictHas('RedisSimple_cache_Date_', toUInt64(1)) on node Stderr: Network roottestconfigdecryption-gw8_default Creating Stderr: Network roottestconfigdecryption-gw8_default Created Stderr: Container roottestconfigdecryption-gw8-node1-1 Creating Stderr: Container roottestconfigdecryption-gw8-node2-1 Creating Stderr: Container roottestconfigdecryption-gw8-node2-1 Created Stderr: Container roottestconfigdecryption-gw8-node1-1 Created Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Stderr: Container roottestconfigdecryption-gw8-node2-1 Starting Stderr: Container roottestconfigdecryption-gw8-node1-1 Starting Stderr: Container roottestconfigdecryption-gw8-node2-1 Started Stderr: Container roottestconfigdecryption-gw8-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.4.3... http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Stdout:0 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Executing query select dictGetString('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Stderr: Container roottestconfigsubstitutions-gw7-node4-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node7-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node8-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node5-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node6-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-node8-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node8-1 Removing Stderr: Container roottestconfigsubstitutions-gw7-node4-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node4-1 Removing Stderr: Container roottestconfigsubstitutions-gw7-node5-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node5-1 Removing Stderr: Container roottestconfigsubstitutions-gw7-node6-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node6-1 Removing Stderr: Container roottestconfigsubstitutions-gw7-node3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node3-1 Removing Stderr: Container roottestconfigsubstitutions-gw7-node2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node2-1 Removing Stderr: Container roottestconfigsubstitutions-gw7-node1-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node1-1 Removing Stderr: Container roottestconfigsubstitutions-gw7-node7-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-node7-1 Removing Stderr: Container roottestconfigsubstitutions-gw7-node8-1 Removed Stderr: Container roottestconfigsubstitutions-gw7-node3-1 Removed Stderr: Container roottestconfigsubstitutions-gw7-zoo3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-zoo1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-zoo2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw7-zoo3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-zoo3-1 Removing Stderr: Container roottestconfigsubstitutions-gw7-zoo1-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-zoo1-1 Removing Stderr: Container roottestconfigsubstitutions-gw7-node6-1 Removed Stderr: Container roottestconfigsubstitutions-gw7-zoo2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw7-zoo2-1 Removing Stderr: Container roottestconfigsubstitutions-gw7-node1-1 Removed Stderr: Container roottestconfigsubstitutions-gw7-node4-1 Removed Stderr: Container roottestconfigsubstitutions-gw7-node5-1 Removed Stderr: Container roottestconfigsubstitutions-gw7-node2-1 Removed Stderr: Container roottestconfigsubstitutions-gw7-node7-1 Removed Stderr: Container roottestconfigsubstitutions-gw7-zoo3-1 Removed Stderr: Container roottestconfigsubstitutions-gw7-zoo1-1 Removed Stderr: Container roottestconfigsubstitutions-gw7-zoo2-1 Removed Stderr: Network roottestconfigsubstitutions-gw7_default Removing Stderr: Network roottestconfigsubstitutions-gw7_default Removed Cleanup called Docker networks for project roottestconfigsubstitutions-gw7 are NETWORK ID NAME DRIVER SCOPE Executing query select dictGetOrDefault('RedisSimple_cache_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Docker containers for project roottestconfigsubstitutions-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigsubstitutions-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigsubstitutions-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select dictGetStringOrDefault('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Unstopped containers: {} No running containers for project: roottestconfigsubstitutions-gw7 Trying to prune unused networks... Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Running tests in /ClickHouse/tests/integration/test_attach_table_normalizer/test.py test_attach_table_normalizer/test.py::test_attach_substr Cluster start called. is_up=False http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query rename table system.text_log to system.text_log_20_test on node2 Docker networks for project roottestattachtablenormalizer-gw7 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachtablenormalizer-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachtablenormalizer-gw7 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestattachtablenormalizer-gw7 are NETWORK ID NAME DRIVER SCOPE Stderr: server Pulling Stderr: server Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/.env --project-name roottestcomposableprotocols-gw6 --file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/server/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/.env --project-name roottestcomposableprotocols-gw6 --file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/server/docker-compose.yml up -d --no-recreate] http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Docker containers for project roottestattachtablenormalizer-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachtablenormalizer-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachtablenormalizer-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select dictGet('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateOrDefault('RedisSimple_cache_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Unstopped containers: {} No running containers for project: roottestattachtablenormalizer-gw7 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:3311 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query rename table system.query_log to system.query_log_20_test on node2 Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/node/database Setup logs dir /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/node/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/.env --project-name roottestattachtablenormalizer-gw7 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/node/docker-compose.yml pull] Stdout:3311 Executing query select 20 on instance http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Executing query select dictGet('RedisSimple_cache_Date_', 'Date_', toUInt64(2)) on node Executing query select dictGetFloat32('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node2-1 bash -c ps -C clickhouse] http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Stdout: PID TTY TIME CMD Stdout: 788 ? 00:00:20 clickhouse run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node2-1 bash -c pkill clickhouse] run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Stdout:788 Executing query select dictGetDate('RedisSimple_cache_Date_', 'Date_', toUInt64(2)) on node Stderr: Network roottestcomposableprotocols-gw6_default Creating Stderr: Network roottestcomposableprotocols-gw6_default Created Stderr: Container roottestcomposableprotocols-gw6-server-1 Creating Stderr: Container roottestcomposableprotocols-gw6-server-1 Created Stderr: Container roottestcomposableprotocols-gw6-server-1 Starting Stderr: Container roottestcomposableprotocols-gw6-server-1 Started ClickHouse instance created get_instance_ip instance_name=server http://localhost:None "GET /v1.46/containers/roottestcomposableprotocols-gw6-server-1/json HTTP/1.1" 200 None get_instance_ip instance_name=server http://localhost:None "GET /v1.46/containers/roottestcomposableprotocols-gw6-server-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in server, ip: 172.16.6.2... http://localhost:None "GET /v1.46/containers/roottestcomposableprotocols-gw6-server-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisSimple_cache_Date_', toUInt64(2)) on node http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select 20 on instance http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7c38fd58f37730ae6db2e2bc77ac1e8cf9da1f45e71670835d50c2bab855f859/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node2-1/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.4.2... http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b90a518439f05d317120e678efe9d6245721ea6e64b1427e23ce34d138d796c1/json HTTP/1.1" 200 None Executing query select dictGetFloat32('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b90a518439f05d317120e678efe9d6245721ea6e64b1427e23ce34d138d796c1/json HTTP/1.1" 200 None ClickHouse node2 started Executing query select value from system.server_settings where name ='max_table_size_to_drop' on node1 http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None test_backup_restore_on_cluster/test.py::test_get_error_from_other_host Executing query CREATE TABLE tbl (`x` UInt8) ENGINE = MergeTree ORDER BY x on node1 Executing query select dictGetFloat32OrDefault('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None Executing query select dictGetDateOrDefault('RedisSimple_cache_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select value from system.server_settings where name ='max_partition_size_to_drop' on node1 Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None Executing query INSERT INTO tbl VALUES (3) on node1 http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select 20 on instance Executing query select dictGet('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Stdout:788 http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None Executing query system reload dictionary RedisSimple_direct_Date_ on node [gw8] PASSED test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml Executing query select value from system.server_settings where name ='max_table_size_to_drop' on node2 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '8') on node1 http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Executing query select dictGetFloat64('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None Executing query select dictGet('RedisSimple_direct_Date_', 'Date_', toUInt64(1)) on node Executing query select value from system.server_settings where name ='max_partition_size_to_drop' on node2 http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://localhost:None "GET /v1.46/containers/e9642f65e01d73b8a719c55eb0e18a701f57a8e3f503bb635961f750a495df4a/json HTTP/1.1" 200 None ClickHouse server started Executing query select dictGetDate('RedisSimple_direct_Date_', 'Date_', toUInt64(1)) on node [gw8] PASSED test_config_decryption/test.py::test_successful_decryption_yaml Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env --project-name roottestconfigdecryption-gw8 --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/docker-compose.yml stop --timeout 20] Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetFloat64('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisSimple_direct_Date_', toUInt64(1)) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DROP DATABASE test_database_backup on instance Executing query select dictGetFloat64OrDefault('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Stdout:788 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_direct_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query DROP DATABASE test_database on instance Executing query select dictGet('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictGetDateOrDefault('RedisSimple_direct_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select dictGetUInt8('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node [gw6] PASSED test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node [gw6] PASSED test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 Executing query select dictGet('RedisSimple_direct_Date_', 'Date_', toUInt64(2)) on node http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_table_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table VALUES (0, 'test_database.test_table'); BACKUP TABLE test_database.test_table TO Disk('backup_disk_local', 'test_table_backup'); CREATE DATABASE test_table_backup ENGINE = Backup('test_database', Disk('backup_disk_local', 'test_table_backup')); on instance Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] Executing query system reload dictionaries on local_node Executing query SYSTEM FLUSH LOGS on server Executing query select dictGetDate('RedisSimple_direct_Date_', 'Date_', toUInt64(2)) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stdout:788 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisSimple_direct_Date_', toUInt64(2)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Executing query select dictGetUInt8OrDefault('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Stdout: PID TTY TIME CMD Stdout: 3311 ? 00:00:03 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] test_backup_restore_on_cluster/test.py::test_keeper_value_max_size Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictGetUInt8('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisSimple_direct_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:3311 Executing query select dictGet('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDateOrDefault('RedisSimple_direct_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SYSTEM FLUSH LOGS on server Executing query INSERT INTO tbl VALUES (111) on node1 Executing query select dictGetUInt16('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query system reload dictionary RedisHash_complex_key_hashed_Date_ on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query INSERT INTO tbl VALUES (222) on node2 Stdout:788 Stdout:1582 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:3311 Executing query select dictGetDate('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node Command:[docker compose --env-file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/.env --project-name roottestcomposableprotocols-gw6 --file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/server/docker-compose.yml stop --timeout 20] [gw6] PASSED test_composable_protocols/test.py::test_proxy_1 Executing query select dictGetUInt16OrDefault('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query SYSTEM STOP REPLICATED SENDS ON CLUSTER 'cluster' tbl on node1 Stderr: Container roottestconfigdecryption-gw8-node2-1 Stopping Stderr: Container roottestconfigdecryption-gw8-node1-1 Stopping Stderr: Container roottestconfigdecryption-gw8-node2-1 Stopped Stderr: Container roottestconfigdecryption-gw8-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env --project-name roottestconfigdecryption-gw8 --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/docker-compose.yml down --volumes] Executing query select dictHas('RedisHash_complex_key_hashed_Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt16('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGet('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query INSERT INTO tbl VALUES (333) on node1 run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw9-node2-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/1c6343f03a37de4eae9384d409f4a6c031d33d71965825a71cd1489d166d52fa/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/1c6343f03a37de4eae9384d409f4a6c031d33d71965825a71cd1489d166d52fa/json HTTP/1.1" 200 586 Executing query select dictGetUInt32('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query INSERT INTO tbl VALUES (444) on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:3311 Stderr: Container roottestconfigdecryption-gw8-node2-1 Stopping Stderr: Container roottestconfigdecryption-gw8-node1-1 Stopping Stderr: Container roottestconfigdecryption-gw8-node2-1 Stopped Stderr: Container roottestconfigdecryption-gw8-node2-1 Removing Stderr: Container roottestconfigdecryption-gw8-node1-1 Stopped Stderr: Container roottestconfigdecryption-gw8-node1-1 Removing Stderr: Container roottestconfigdecryption-gw8-node2-1 Removed Stderr: Container roottestconfigdecryption-gw8-node1-1 Removed Stderr: Network roottestconfigdecryption-gw8_default Removing Stderr: Network roottestconfigdecryption-gw8_default Removed Cleanup called Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Docker networks for project roottestconfigdecryption-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfigdecryption-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigdecryption-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigdecryption-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfigdecryption-gw8 Trying to prune unused networks... Executing query select dictGetDateOrDefault('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '9') on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 133f11833cd2 ENV SHLVL 0 ENV HOME /root ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV UBSAN_OPTIONS print_stacktrace=1 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_aggregation_memory_efficient/test.py::test_remote test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile test_analyzer_compatibility/test.py::test_two_new_versions test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables test_attach_table_normalizer/test.py::test_attach_substr test_attach_table_normalizer/test.py::test_attach_substr_restart test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 test_config_corresponding_root/test.py::test_work test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed test_config_substitutions/test.py::test_allow_databases test_config_substitutions/test.py::test_config test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_local'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[File('"'"'test_database_backup_file'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_local'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[File('"'"'test_table_backup_file'"'"')]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Date_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[String_]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV CLICKHOUSE_USE_OLD_ANALYZER 1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_BASE_TAG 8b2301119731 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 ENV WORKER_FREE_PORTS 30400 30401 30402 30403 30404 30405 30406 30407 30408 30409 30410 30411 30412 30413 30414 30415 30416 30417 30418 30419 30420 30421 30422 30423 30424 30425 30426 30427 30428 30429 30430 30431 30432 30433 30434 30435 30436 30437 30438 30439 30440 30441 30442 30443 30444 30445 30446 30447 30448 30449 ENV PYTEST_XDIST_TESTRUNUID 403ebba7b09548d7ae9e72b89a600d03 ENV PYTEST_XDIST_WORKER gw8 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication (setup) CLUSTER INIT base_config_dir:/clickhouse-config clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Setup Keeper Cluster name: project_name:roottestazureblobstoragezerocopyreplication-gw8. Added instance name:node1 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env', '--project-name', 'roottestazureblobstoragezerocopyreplication-gw8', '--file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name: project_name:roottestazureblobstoragezerocopyreplication-gw8. Added instance name:node2 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env', '--project-name', 'roottestazureblobstoragezerocopyreplication-gw8', '--file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml', '--file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Starting cluster... Running tests in /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/test.py Cluster start called. is_up=False Docker networks for project roottestazureblobstoragezerocopyreplication-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestazureblobstoragezerocopyreplication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestazureblobstoragezerocopyreplication-gw8 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestazureblobstoragezerocopyreplication-gw8 are NETWORK ID NAME DRIVER SCOPE Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Docker containers for project roottestazureblobstoragezerocopyreplication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestazureblobstoragezerocopyreplication-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestazureblobstoragezerocopyreplication-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select dictGet('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Unstopped containers: {} No running containers for project: roottestazureblobstoragezerocopyreplication-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/configs/config.d/config.xml', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_gen/storage_conf.xml'] to /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/database Setup logs dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/configs/config.d/config.xml', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_gen/storage_conf.xml'] to /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/database Setup logs dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/coordination', 'AZURITE_PORT': '30400', 'AZURITE_STORAGE_ACCOUNT_URL': 'http://azurite1:30400/devstoreaccount1', 'AZURITE_CONNECTION_STRING': 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30400/devstoreaccount1;'} stored in /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --project-name roottestazureblobstoragezerocopyreplication-gw8 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml pull] Stderr: Container roottestcomposableprotocols-gw6-server-1 Stopping Stderr: Container roottestcomposableprotocols-gw6-server-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/server/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/server/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/.env --project-name roottestcomposableprotocols-gw6 --file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw6/server/docker-compose.yml down --volumes] Executing query select dictGetUInt32OrDefault('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query select dictGetDate('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1619 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1619 Executing query select 20 on node2 Executing query select dictGet('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisHash_complex_key_hashed_Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt32('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:3311 Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Stderr: Container roottestcomposableprotocols-gw6-server-1 Stopping Stderr: Container roottestcomposableprotocols-gw6-server-1 Stopped Stderr: Container roottestcomposableprotocols-gw6-server-1 Removing Stderr: Container roottestcomposableprotocols-gw6-server-1 Removed Stderr: Network roottestcomposableprotocols-gw6_default Removing Stderr: Network roottestcomposableprotocols-gw6_default Removed Cleanup called Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUInt64('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Docker networks for project roottestcomposableprotocols-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcomposableprotocols-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcomposableprotocols-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcomposableprotocols-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestcomposableprotocols-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Executing query select dictGetDateOrDefault('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '9') on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select 20 on node2 Executing query system reload dictionary RedisHash_complex_key_cache_Date_ on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query select dictGet('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt64OrDefault('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:3311 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictGetDate('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on node2 Executing query select dictGet('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt64('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisHash_complex_key_cache_Date_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetInt8('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.text_log_18_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.query_log_18_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query select dictGetDateOrDefault('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select count() from system.text_log_5_test on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query select dictGet('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "POST /v1.46/exec/3c74b39bf0a5669731e4c0b005877fec6c515954b26dd30080bd41573193eae9/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/3c74b39bf0a5669731e4c0b005877fec6c515954b26dd30080bd41573193eae9/json HTTP/1.1" 200 586 Executing query select count() from system.query_log_19_test on node2 Executing query select dictGetInt8OrDefault('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetDate('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.text_log_11_test on node2 Executing query select dictGet('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetInt8('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisHash_complex_key_cache_Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt16('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.query_log_8_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/.env --project-name roottestattachtablenormalizer-gw7 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/.env --project-name roottestattachtablenormalizer-gw7 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/node/docker-compose.yml up -d --no-recreate] Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:4147 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select count() from system.asynchronous_loader where job ilike '%_log_%_test' and execution_pool = 'BackgroundLoad' on node2 Stdout:4147 Executing query select 20 on instance Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query select dictGetDateOrDefault('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query system flush logs on node2 Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query select dictGetInt16OrDefault('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query system reload dictionary RedisHash_complex_key_direct_Date_ on node test_backup_restore_on_cluster/test.py::test_mutation Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY tuple() on node1 Stderr: Network roottestattachtablenormalizer-gw7_default Creating Stderr: Network roottestattachtablenormalizer-gw7_default Created Stderr: Container roottestattachtablenormalizer-gw7-node-1 Creating Stderr: Container roottestattachtablenormalizer-gw7-node-1 Created Stderr: Container roottestattachtablenormalizer-gw7-node-1 Starting Stderr: Container roottestattachtablenormalizer-gw7-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestattachtablenormalizer-gw7-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestattachtablenormalizer-gw7-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.4.2... http://localhost:None "GET /v1.46/containers/roottestattachtablenormalizer-gw7-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query select dictGet('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on instance Executing query select dictGet('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query select dictGetDate('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(5) on node1 http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query select dictGetInt32('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query select dictGetInt16('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query select dictHas('RedisHash_complex_key_direct_Date_', tuple(toUInt64(1),toString('world'))) on node Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(5, 5) on node2 http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query select 20 on instance Executing query rename table system.text_log to system.text_log_21_test on node2 http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(10, 5) on node1 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query rename table system.query_log to system.query_log_21_test on node2 Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query select dictGetDateOrDefault('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query system flush logs on node2 Executing query ALTER TABLE tbl UPDATE x=x+1 WHERE 1 on node1 Executing query select dictGetInt32OrDefault('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query DROP DATABASE test_table_backup on instance Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query select dictGet('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None Executing query ALTER TABLE tbl UPDATE x=x+1+sleep(3) WHERE 1 on node1 Executing query select dictGet('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP DATABASE test_database on instance http://localhost:None "GET /v1.46/containers/3ebdc75061501a70e6ea46e23c0d34ac73becd26974ed39244859f61ed37b559/json HTTP/1.1" 200 None ClickHouse node started Executing query DROP TABLE IF EXISTS default.file on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env --project-name roottestclusterallreplicas-gw4 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/docker-compose.yml stop --timeout 20] [gw4] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetDate('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query CREATE TABLE default.file(`s` String, `n` UInt8) ENGINE = MergeTree PARTITION BY substring(s, 1, 2) ORDER BY n on node Executing query ALTER TABLE tbl UPDATE x=x+1+sleep(3) WHERE 1 on node1 Executing query select dictGetInt64('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query select dictGetInt32('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictHas('RedisHash_complex_key_direct_Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DETACH TABLE file on node http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_table_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table VALUES (0, 'test_database.test_table'); BACKUP TABLE test_database.test_table TO Disk('backup_disk_object_storage_local_plain', 'test_table_backup'); CREATE DATABASE test_table_backup ENGINE = Backup('test_database', Disk('backup_disk_object_storage_local_plain', 'test_table_backup')); on instance Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '10') on node1 Executing query rename table system.text_log to system.text_log_22_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node run container_id:roottestattachtablenormalizer-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql"] Command:[docker exec -u root roottestattachtablenormalizer-gw7-node-1 bash -c sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql] Executing query ATTACH TABLE file on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query rename table system.query_log to system.query_log_22_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query select dictGetDateOrDefault('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node [gw7] PASSED test_attach_table_normalizer/test.py::test_attach_substr Executing query DROP TABLE IF EXISTS default.file on node test_attach_table_normalizer/test.py::test_attach_substr_restart Executing query system flush logs on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Executing query select dictGetInt64OrDefault('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Stdout: PID TTY TIME CMD Stdout: 4147 ? 00:00:03 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query CREATE TABLE default.file(`s` String, `n` UInt8) ENGINE = MergeTree PARTITION BY substring(s, 1, 2) ORDER BY n on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Stdout:4147 Executing query system reload dictionaries on node Executing query select dictGet('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestattachtablenormalizer-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql"] Command:[docker exec -u root roottestattachtablenormalizer-gw7-node-1 bash -c sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql] [gw2] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] Run test with id: 12 Executing query system reload dictionaries on node run container_id:roottestattachtablenormalizer-gw7-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestattachtablenormalizer-gw7-node-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:01 clickhouse run container_id:roottestattachtablenormalizer-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill -9 clickhouse'] Command:[docker exec -u root roottestattachtablenormalizer-gw7-node-1 bash -c pkill -9 clickhouse] run container_id:roottestattachtablenormalizer-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt64('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUUID('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Stdout:8 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query system reload dictionary RedisSimple_flat_Float32_ on node Executing query select dictGet('RedisSimple_flat_Float32_', 'Float32_', toUInt64(1)) on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query rename table system.text_log to system.text_log_23_test on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetFloat32('RedisSimple_flat_Float32_', 'Float32_', toUInt64(1)) on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Stdout:4147 Executing query rename table system.query_log to system.query_log_23_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query select dictGetUUIDOrDefault('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query system flush logs on node2 Executing query select dictHas('RedisSimple_flat_Float32_', toUInt64(1)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node run container_id:roottestattachtablenormalizer-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestattachtablenormalizer-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestattachtablenormalizer-gw7-node-1/exec HTTP/1.1" 201 74 Executing query select dictGetOrDefault('RedisSimple_flat_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node http://localhost:None "POST /v1.46/exec/a73ee5d4bc6e767e0ec0e4e9ca2b58d182b098660e7b32a518c5d2fd39c87ef2/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/a73ee5d4bc6e767e0ec0e4e9ca2b58d182b098660e7b32a518c5d2fd39c87ef2/json HTTP/1.1" 200 586 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetDate('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat32OrDefault('RedisSimple_flat_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:4147 Executing query select dictGetUUID('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('RedisSimple_flat_Float32_', 'Float32_', toUInt64(2)) on node Executing query rename table system.text_log to system.text_log_24_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat32('RedisSimple_flat_Float32_', 'Float32_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '10') on node1 Executing query rename table system.query_log to system.query_log_24_test on node2 run container_id:roottestattachtablenormalizer-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Stdout:739 Clickhouse process running. run container_id:roottestattachtablenormalizer-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('RedisSimple_flat_Float32_', toUInt64(2)) on node Executing query select dictGetDateOrDefault('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query system flush logs on node2 Stdout:739 Executing query select 20 on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:4147 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_mutation Executing query select dictGetFloat32OrDefault('RedisSimple_flat_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetDateTime('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select 20 on node Executing query system reload dictionary RedisSimple_hashed_Float32_ on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetDate('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/.env --project-name roottestattachtablenormalizer-gw7 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/node/docker-compose.yml stop --timeout 20] [gw7] PASSED test_attach_table_normalizer/test.py::test_attach_substr_restart Executing query select dictGet('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(1)) on node Executing query rename table system.text_log to system.text_log_25_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetFloat32('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(1)) on node Executing query rename table system.query_log to system.query_log_25_test on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Stdout:4147 Stderr: Container roottestattachtablenormalizer-gw7-node-1 Stopping Stderr: Container roottestattachtablenormalizer-gw7-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/.env --project-name roottestattachtablenormalizer-gw7 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw7/node/docker-compose.yml down --volumes] Executing query select dictHas('RedisSimple_hashed_Float32_', toUInt64(1)) on node Executing query system flush logs on node2 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGetOrDefault('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetString('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat32OrDefault('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Stderr: Container roottestattachtablenormalizer-gw7-node-1 Stopping Stderr: Container roottestattachtablenormalizer-gw7-node-1 Stopped Stderr: Container roottestattachtablenormalizer-gw7-node-1 Removing Stderr: Container roottestattachtablenormalizer-gw7-node-1 Removed Stderr: Network roottestattachtablenormalizer-gw7_default Removing Stderr: Network roottestattachtablenormalizer-gw7_default Removed Cleanup called Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Docker networks for project roottestattachtablenormalizer-gw7 are NETWORK ID NAME DRIVER SCOPE test_backup_restore_on_cluster/test.py::test_projection Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}') ORDER BY y PARTITION BY x%10 on node1 Docker containers for project roottestattachtablenormalizer-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachtablenormalizer-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachtablenormalizer-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select dictGetDateTime('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node Unstopped containers: {} No running containers for project: roottestattachtablenormalizer-gw7 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:Total reclaimed space: 0B Volumes pruned: 5 test_config_corresponding_root/test.py::test_work Running tests in /ClickHouse/tests/integration/test_config_corresponding_root/test.py Cluster start called. is_up=False Executing query select dictGet('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(2)) on node Docker networks for project roottestconfigcorrespondingroot-gw7 are NETWORK ID NAME DRIVER SCOPE No clickhouse process running. Start new one. Executing query rename table system.text_log to system.text_log_26_test on node2 Docker containers for project roottestconfigcorrespondingroot-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 Docker volumes for project roottestconfigcorrespondingroot-gw7 are DRIVER VOLUME NAME Cleanup called http://localhost:None "POST /v1.46/exec/ea3cccbad5f43ed5ab3c1b3d84f50a8bac873ca3519d4d425890f3c9b9b2285c/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/ea3cccbad5f43ed5ab3c1b3d84f50a8bac873ca3519d4d425890f3c9b9b2285c/json HTTP/1.1" 200 586 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Docker networks for project roottestconfigcorrespondingroot-gw7 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfigcorrespondingroot-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigcorrespondingroot-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigcorrespondingroot-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfigcorrespondingroot-gw7 Trying to prune unused networks... Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Executing query select dictGetFloat32('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(2)) on node Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_corresponding_root/configs/config.d/bad.xml'] to /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/node/database Setup logs dir /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(3) on node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/.env --project-name roottestconfigcorrespondingroot-gw7 --file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/node/docker-compose.yml pull] Executing query rename table system.query_log to system.query_log_26_test on node2 Executing query select dictGetStringOrDefault('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query system flush logs on node2 Executing query select dictHas('RedisSimple_hashed_Float32_', toUInt64(2)) on node Executing query ALTER TABLE tbl ADD PROJECTION prjmax (SELECT MAX(x)) on node1 Executing query select dictGet('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query INSERT INTO tbl VALUES (100, 'a'), (101, 'b') on node1 Executing query select dictGetFloat32('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat32OrDefault('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT count() FROM system.projection_parts WHERE database='default' AND table='tbl' AND name='prjmax' on node1 Stdout:4991 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Stdout:4991 Executing query select 20 on instance Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query system reload dictionary RedisSimple_cache_Float32_ on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGetString('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '11') on node1 Executing query rename table system.text_log to system.text_log_27_test on node2 Executing query select dictGet('RedisSimple_cache_Float32_', 'Float32_', toUInt64(1)) on node Executing query select dictGetFloat32OrDefault('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query rename table system.query_log to system.query_log_27_test on node2 Executing query select dictGetFloat32('RedisSimple_cache_Float32_', 'Float32_', toUInt64(1)) on node Executing query select 20 on instance Executing query select dictGet('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query system flush logs on node2 Executing query select dictHas('RedisSimple_cache_Float32_', toUInt64(1)) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetFloat64('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT count() FROM system.projection_parts WHERE database='default' AND table='tbl' AND name='prjmax' on node1 Executing query select 20 on instance Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat32OrDefault('RedisSimple_cache_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query RESTORE TABLE tbl FROM Disk('backups', '11') on node1 Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query select dictGetFloat64OrDefault('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query rename table system.text_log to system.text_log_28_test on node2 Executing query select dictGetFloat32('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGet('RedisSimple_cache_Float32_', 'Float32_', toUInt64(2)) on node Executing query DROP DATABASE test_table_backup on instance [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] Executing query system reload dictionaries on node Executing query rename table system.query_log to system.query_log_28_test on node2 Executing query select dictGetFloat32('RedisSimple_cache_Float32_', 'Float32_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGet('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP DATABASE test_database on instance Executing query system flush logs on node2 Executing query select dictHas('RedisSimple_cache_Float32_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query SELECT count() FROM system.projection_parts WHERE database='default' AND table='tbl' AND name='prjmax' on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query select dictGetUInt8('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select dictGetOrDefault('RedisSimple_cache_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_table_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table VALUES (0, 'test_database.test_table'); BACKUP TABLE test_database.test_table TO Disk('backup_disk_s3_plain', 'test_table_backup'); CREATE DATABASE test_table_backup ENGINE = Backup('test_database', Disk('backup_disk_s3_plain', 'test_table_backup')); on instance [gw0] PASSED test_backup_restore_on_cluster/test.py::test_projection Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32OrDefault('RedisSimple_cache_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query system reload dictionary RedisSimple_direct_Float32_ on node Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query rename table system.text_log to system.text_log_29_test on node2 Executing query select dictGetFloat64('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUInt8OrDefault('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictGet('RedisSimple_direct_Float32_', 'Float32_', toUInt64(1)) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query rename table system.query_log to system.query_log_29_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Executing query select dictGet('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Stdout: PID TTY TIME CMD Stdout: 4991 ? 00:00:06 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] Executing query select dictGetFloat32('RedisSimple_direct_Float32_', 'Float32_', toUInt64(1)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:4991 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query system flush logs on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictGetUInt16('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisSimple_direct_Float32_', toUInt64(1)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisSimple_direct_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetFloat32OrDefault('RedisSimple_direct_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node test_backup_restore_on_cluster/test.py::test_replicated_database Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt16OrDefault('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Stdout:4991 Executing query select dictGet('RedisSimple_direct_Float32_', 'Float32_', toUInt64(2)) on node Executing query select dictGetUInt8('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query CREATE TABLE mydb.tbl(x UInt8, y String) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Executing query rename table system.text_log to system.text_log_30_test on node2 Executing query select dictGet('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32('RedisSimple_direct_Float32_', 'Float32_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query rename table system.query_log to system.query_log_30_test on node2 Executing query select dictGetUInt32('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisSimple_direct_Float32_', toUInt64(2)) on node Executing query SYSTEM SYNC DATABASE REPLICA mydb on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node2-1 bash -c ps -C clickhouse] Executing query select dictGetOrDefault('RedisSimple_direct_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Stdout: PID TTY TIME CMD Stdout: 1619 ? 00:00:23 clickhouse run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node2-1 bash -c pkill clickhouse] Executing query INSERT INTO mydb.tbl VALUES (1, 'Don''t') on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1619 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetFloat32OrDefault('RedisSimple_direct_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Stdout:4991 Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query INSERT INTO mydb.tbl VALUES (2, 'count') on node2 Stderr: Container roottestclusterallreplicas-gw4-node2-1 Stopping Stderr: Container roottestclusterallreplicas-gw4-node1-1 Stopping Stderr: Container roottestclusterallreplicas-gw4-node2-1 Stopped Stderr: Container roottestclusterallreplicas-gw4-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env --project-name roottestclusterallreplicas-gw4 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/docker-compose.yml down --volumes] Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query system reload dictionary RedisHash_complex_key_hashed_Float32_ on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query INSERT INTO mydb.tbl VALUES (3, 'your') on node1 Executing query select dictGet('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt32OrDefault('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query select dictGetUInt16('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query INSERT INTO mydb.tbl VALUES (4, 'chickens') on node2 Executing query select dictGetFloat32('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Stderr: Container roottestclusterallreplicas-gw4-node1-1 Stopping Stderr: Container roottestclusterallreplicas-gw4-node2-1 Stopping Stderr: Container roottestclusterallreplicas-gw4-node2-1 Stopped Stderr: Container roottestclusterallreplicas-gw4-node2-1 Removing Stderr: Container roottestclusterallreplicas-gw4-node1-1 Stopped Stderr: Container roottestclusterallreplicas-gw4-node1-1 Removing Stderr: Container roottestclusterallreplicas-gw4-node2-1 Removed Stderr: Container roottestclusterallreplicas-gw4-node1-1 Removed Stderr: Network roottestclusterallreplicas-gw4_default Removing Stderr: Network roottestclusterallreplicas-gw4_default Removed Cleanup called Docker networks for project roottestclusterallreplicas-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestclusterallreplicas-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query select dictGet('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node Docker volumes for project roottestclusterallreplicas-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestclusterallreplicas-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Unstopped containers: {} No running containers for project: roottestclusterallreplicas-gw4 Trying to prune unused networks... Executing query OPTIMIZE TABLE mydb.tbl ON CLUSTER 'cluster' FINAL on node1 Trying to prune unused images... Command:[docker image prune -f] Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stdout:1619 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 test_aggregation_memory_efficient/test.py::test_remote Running tests in /ClickHouse/tests/integration/test_aggregation_memory_efficient/test.py Cluster start called. is_up=False Docker networks for project roottestaggregationmemoryefficient-gw4 are NETWORK ID NAME DRIVER SCOPE run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Docker containers for project roottestaggregationmemoryefficient-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query select dictHas('RedisHash_complex_key_hashed_Float32_', tuple(toUInt64(1),toString('world'))) on node Docker volumes for project roottestaggregationmemoryefficient-gw4 are DRIVER VOLUME NAME Cleanup called Stdout:4991 Docker networks for project roottestaggregationmemoryefficient-gw4 are NETWORK ID NAME DRIVER SCOPE Executing query select dictGetUInt64('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node Docker containers for project roottestaggregationmemoryefficient-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestaggregationmemoryefficient-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestaggregationmemoryefficient-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestaggregationmemoryefficient-gw4 Trying to prune unused networks... Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/database Setup logs dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/database Setup logs dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env --project-name roottestaggregationmemoryefficient-gw4 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/docker-compose.yml pull] Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt64OrDefault('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Stdout:1619 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt32('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetFloat32('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/9bc9c307c531f98d8da3afe908db35d3e133d3039afc2e9d914b26420de3ae2f/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/9bc9c307c531f98d8da3afe908db35d3e133d3039afc2e9d914b26420de3ae2f/json HTTP/1.1" 200 586 Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '12') SETTINGS replica_num=2 on node1 Executing query select dictGet('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisHash_complex_key_hashed_Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt8('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1619 Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '12') on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query system reload dictionary RedisHash_complex_key_cache_Float32_ on node Stdout:5822 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stdout:5822 Executing query select 20 on instance Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/.env --project-name roottestconfigcorrespondingroot-gw7 --file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/.env --project-name roottestconfigcorrespondingroot-gw7 --file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/node/docker-compose.yml up -d --no-recreate] Executing query select dictGet('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt8OrDefault('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Stderr: node1 Skipped - Image is already being pulled by node2 Stderr: node2 Pulling Stderr: node2 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env --project-name roottestaggregationmemoryefficient-gw4 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env --project-name roottestaggregationmemoryefficient-gw4 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate] Executing query select dictGetUInt64('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetFloat32('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('RedisHash_complex_key_cache_Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on instance Stdout:1619 Executing query select dictGetInt16('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Stderr: Network roottestconfigcorrespondingroot-gw7_default Creating Stderr: Network roottestconfigcorrespondingroot-gw7_default Created Stderr: Container roottestconfigcorrespondingroot-gw7-node-1 Creating Stderr: Container roottestconfigcorrespondingroot-gw7-node-1 Created Stderr: Container roottestconfigcorrespondingroot-gw7-node-1 Starting Stderr: Container roottestconfigcorrespondingroot-gw7-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestconfigcorrespondingroot-gw7-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestconfigcorrespondingroot-gw7-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.2.2... http://localhost:None "GET /v1.46/containers/roottestconfigcorrespondingroot-gw7-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b515ae3a45423cf4db1297af7983e6ff7b1459f5713a0e7a83dd2c431298883a/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b515ae3a45423cf4db1297af7983e6ff7b1459f5713a0e7a83dd2c431298883a/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 http://localhost:None "GET /v1.46/containers/b515ae3a45423cf4db1297af7983e6ff7b1459f5713a0e7a83dd2c431298883a/json HTTP/1.1" 200 None Stderr: Network roottestaggregationmemoryefficient-gw4_default Creating Stderr: Network roottestaggregationmemoryefficient-gw4_default Created Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Creating Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Creating Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Created Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Created Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Starting Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Starting Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Started Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.4.3... http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b515ae3a45423cf4db1297af7983e6ff7b1459f5713a0e7a83dd2c431298883a/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b515ae3a45423cf4db1297af7983e6ff7b1459f5713a0e7a83dd2c431298883a/json HTTP/1.1" 200 None Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None Executing query select 20 on instance http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b515ae3a45423cf4db1297af7983e6ff7b1459f5713a0e7a83dd2c431298883a/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b515ae3a45423cf4db1297af7983e6ff7b1459f5713a0e7a83dd2c431298883a/logs?stderr=1&stdout=1×tamps=0&follow=0&tail=all HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b515ae3a45423cf4db1297af7983e6ff7b1459f5713a0e7a83dd2c431298883a/json HTTP/1.1" 200 None Failed to start cluster: Instance `node' failed to start. Container status: exited, logs: Processing configuration file '/etc/clickhouse-server/config.xml'. Merging configuration file '/etc/clickhouse-server/conf.d/macros.xml'. Merging configuration file '/etc/clickhouse-server/config.d/0_common_enable_keeper_async_replication.xml'. Merging configuration file '/etc/clickhouse-server/config.d/0_common_instance_config.xml'. Merging configuration file '/etc/clickhouse-server/config.d/bad.xml'. Poco::Exception. Code: 1000, e.code() = 0, Exception: Failed to merge config with '/etc/clickhouse-server/config.d/bad.xml': Exception: Root element doesn't have the corresponding root element as the config file. It must be , Stack trace (when copying this message, always include the lines below): 0. ./contrib/llvm-project/libcxx/include/__exception/exception.h:106: Poco::Exception::Exception(String&&, int) @ 0x0000000036e1adf1 1. ./build_docker/./src/Common/Config/ConfigProcessor.cpp:729: DB::ConfigProcessor::processConfig(bool*, zkutil::ZooKeeperNodeCache*, std::shared_ptr const&, bool) @ 0x0000000030643a64 2. ./build_docker/./src/Common/Config/ConfigProcessor.cpp:802: DB::ConfigProcessor::loadConfig(bool, bool) @ 0x0000000030644126 3. ./build_docker/./src/Daemon/BaseDaemon.cpp:126: BaseDaemon::reloadConfiguration() @ 0x000000001b72da4a 4. ./build_docker/./src/Daemon/BaseDaemon.cpp:255: BaseDaemon::initialize(Poco::Util::Application&) @ 0x000000001b7306ee 5. ./build_docker/./programs/server/Server.cpp:579: DB::Server::initialize(Poco::Util::Application&) @ 0x000000001b5d65b2 6. ./build_docker/./base/poco/Util/src/Application.cpp:310: Poco::Util::Application::run() @ 0x0000000037058945 7. ./build_docker/./programs/server/Server.cpp:573: DB::Server::run() @ 0x000000001b5d61da 8. ./build_docker/./programs/server/Server.cpp:371: mainEntryClickHouseServer(int, char**) @ 0x000000001b5d1339 9. ./build_docker/./programs/main.cpp:295: main @ 0x000000000b6e089f 10. ? @ 0x00007ff89a7f2d90 11. ? @ 0x00007ff89a7f2e40 12. _start @ 0x000000000b60902e (version 25.2.2.173926.altinityantalya.173926 (official build)) Traceback (most recent call last): File "/ClickHouse/tests/integration/helpers/cluster.py", line 3085, in start instance.wait_for_start(start_timeout) File "/ClickHouse/tests/integration/helpers/cluster.py", line 4386, in wait_for_start raise Exception( Exception: Instance `node' failed to start. Container status: exited, logs: Processing configuration file '/etc/clickhouse-server/config.xml'. Merging configuration file '/etc/clickhouse-server/conf.d/macros.xml'. Merging configuration file '/etc/clickhouse-server/config.d/0_common_enable_keeper_async_replication.xml'. Merging configuration file '/etc/clickhouse-server/config.d/0_common_instance_config.xml'. Merging configuration file '/etc/clickhouse-server/config.d/bad.xml'. Poco::Exception. Code: 1000, e.code() = 0, Exception: Failed to merge config with '/etc/clickhouse-server/config.d/bad.xml': Exception: Root element doesn't have the corresponding root element as the config file. It must be , Stack trace (when copying this message, always include the lines below): 0. ./contrib/llvm-project/libcxx/include/__exception/exception.h:106: Poco::Exception::Exception(String&&, int) @ 0x0000000036e1adf1 1. ./build_docker/./src/Common/Config/ConfigProcessor.cpp:729: DB::ConfigProcessor::processConfig(bool*, zkutil::ZooKeeperNodeCache*, std::shared_ptr const&, bool) @ 0x0000000030643a64 2. ./build_docker/./src/Common/Config/ConfigProcessor.cpp:802: DB::ConfigProcessor::loadConfig(bool, bool) @ 0x0000000030644126 3. ./build_docker/./src/Daemon/BaseDaemon.cpp:126: BaseDaemon::reloadConfiguration() @ 0x000000001b72da4a 4. ./build_docker/./src/Daemon/BaseDaemon.cpp:255: BaseDaemon::initialize(Poco::Util::Application&) @ 0x000000001b7306ee 5. ./build_docker/./programs/server/Server.cpp:579: DB::Server::initialize(Poco::Util::Application&) @ 0x000000001b5d65b2 6. ./build_docker/./base/poco/Util/src/Application.cpp:310: Poco::Util::Application::run() @ 0x0000000037058945 7. ./build_docker/./programs/server/Server.cpp:573: DB::Server::run() @ 0x000000001b5d61da 8. ./build_docker/./programs/server/Server.cpp:371: mainEntryClickHouseServer(int, char**) @ 0x000000001b5d1339 9. ./build_docker/./programs/main.cpp:295: main @ 0x000000000b6e089f 10. ? @ 0x00007ff89a7f2d90 11. ? @ 0x00007ff89a7f2e40 12. _start @ 0x000000000b60902e (version 25.2.2.173926.altinityantalya.173926 (official build)) Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/.env --project-name roottestconfigcorrespondingroot-gw7 --file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/node/docker-compose.yml stop --timeout 20] http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database Stderr: Container roottestconfigcorrespondingroot-gw7-node-1 Stopping Stderr: Container roottestconfigcorrespondingroot-gw7-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/.env --project-name roottestconfigcorrespondingroot-gw7 --file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw7/node/docker-compose.yml down --volumes] Executing query SELECT id, value FROM test_table_backup.test_table on instance http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw9-node2-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/078b712937eb486dbaf8277612a9460b8a1382c1e4f5b19819778e8e7cce156b/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/078b712937eb486dbaf8277612a9460b8a1382c1e4f5b19819778e8e7cce156b/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetInt16OrDefault('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetFloat32('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt8('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP DATABASE test_table_backup on instance http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None Stderr: Container roottestconfigcorrespondingroot-gw7-node-1 Stopping Stderr: Container roottestconfigcorrespondingroot-gw7-node-1 Stopped Stderr: Container roottestconfigcorrespondingroot-gw7-node-1 Removing Stderr: Container roottestconfigcorrespondingroot-gw7-node-1 Removed Stderr: Network roottestconfigcorrespondingroot-gw7_default Removing Stderr: Network roottestconfigcorrespondingroot-gw7_default Removed Cleanup called Executing query select dictGet('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisHash_complex_key_cache_Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Docker networks for project roottestconfigcorrespondingroot-gw7 are NETWORK ID NAME DRIVER SCOPE http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None Docker containers for project roottestconfigcorrespondingroot-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query DROP DATABASE test_database on instance Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Docker volumes for project roottestconfigcorrespondingroot-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigcorrespondingroot-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfigcorrespondingroot-gw7 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 [gw7] PASSED test_config_corresponding_root/test.py::test_work http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query select dictGetInt32('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node http://localhost:None "GET /v1.46/containers/6d74ac47c075550702a7e50c6671e8057aacb8146ef2f9f1605ee7ed3b51b69b/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.4.2... http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/25d3d41f9965a95919a94c89c66c9d33eff901bb51b701b7e0cb32d3f3115201/json HTTP/1.1" 200 None ClickHouse node2 started Executing query create table da_memory_efficient_shard(A Int64, B Int64) Engine=MergeTree order by A partition by B % 2; on node1 http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/.backup HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/checksums.txt HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/columns.txt HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/count.txt HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/data.bin HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/data.cmrk3 HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/default_compression_codec.txt HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/metadata_version.txt HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/primary.cidx HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/serialization.json HTTP/1.1" 204 0 http://172.16.7.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/metadata/test_database/test_table.sql HTTP/1.1" 204 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_table_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table VALUES (0, 'test_database.test_table'); BACKUP TABLE test_database.test_table TO File('test_table_backup_file'); CREATE DATABASE test_table_backup ENGINE = Backup('test_database', File('test_table_backup_file')); on instance Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Stdout:2463 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query create table da_memory_efficient_shard(A Int64, B Int64) Engine=MergeTree order by A partition by B % 2; on node2 Stdout:2463 Executing query select 20 on node2 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query system reload dictionary RedisHash_complex_key_direct_Float32_ on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query insert into da_memory_efficient_shard select number, number from numbers(100000); on node1 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt32OrDefault('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] test_backup_restore_on_cluster/test.py::test_replicated_database_async Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Stdout: PID TTY TIME CMD Stdout: 5822 ? 00:00:04 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] Executing query select dictGetInt16('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:5822 Executing query insert into da_memory_efficient_shard select number + 100000, number from numbers(100000); on node2 Executing query select dictGetFloat32('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on node2 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query CREATE TABLE mydb.tbl(x UInt8) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Executing query select dictHas('RedisHash_complex_key_direct_Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query set distributed_aggregation_memory_efficient = 1, group_by_two_level_threshold = 1, group_by_two_level_threshold_bytes=1 on node1 Executing query select dictGetInt64('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query select sum(a) from (SELECT B, uniqExact(A) a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY B) on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query CREATE TABLE mydb.tbl2(y String) ENGINE=ReplicatedMergeTree ORDER BY y on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on node2 Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SYSTEM SYNC DATABASE REPLICA mydb on node2 Stdout:5822 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.text_log_25_test on node2 Executing query select dictGetInt64OrDefault('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query set distributed_aggregation_memory_efficient = 0 on node1 Executing query INSERT INTO mydb.tbl VALUES (1) on node1 Executing query select dictGetInt32('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select sum(a) from (SELECT B, uniqExact(A) a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY B) on node1 Executing query select dictGetFloat32('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.query_log_25_test on node2 Executing query INSERT INTO mydb.tbl VALUES (22) on node1 Executing query select dictHas('RedisHash_complex_key_direct_Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetUUID('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.text_log_27_test on node2 Executing query INSERT INTO mydb.tbl2 VALUES ('a') on node2 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query set distributed_aggregation_memory_efficient = 1, group_by_two_level_threshold = 1, group_by_two_level_threshold_bytes=1 on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query INSERT INTO mydb.tbl2 VALUES ('bb') on node2 Executing query select count() from system.query_log_9_test on node2 Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Stdout:5822 Executing query SELECT fullHostName() AS h, uniqExact(A) AS a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY h ORDER BY h; on node1 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query system reload dictionaries on node Executing query OPTIMIZE TABLE mydb.tbl ON CLUSTER 'cluster' FINAL on node1 Executing query select dictGetUUIDOrDefault('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select count() from system.text_log_15_test on node2 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query set distributed_aggregation_memory_efficient = 0 on node1 Executing query select dictGet('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.query_log_7_test on node2 [gw2] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] Run test with id: 13 Executing query system reload dictionaries on node Executing query select dictGetInt64('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SELECT fullHostName() AS h, uniqExact(A) AS a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY h ORDER BY h; on node1 Executing query OPTIMIZE TABLE mydb.tbl2 ON CLUSTER 'cluster' FINAL on node1 Executing query system reload dictionary RedisSimple_flat_Float64_ on node Executing query select dictGetDate('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.asynchronous_loader where job ilike '%_log_%_test' and execution_pool = 'BackgroundLoad' on node2 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('RedisSimple_flat_Float64_', 'Float64_', toUInt64(1)) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env --project-name roottestaggregationmemoryefficient-gw4 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/docker-compose.yml stop --timeout 20] [gw4] PASSED test_aggregation_memory_efficient/test.py::test_remote No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node http://localhost:None "POST /v1.46/exec/1901fa6df621cf9101825636b9ce07df167541bf169b515a170559ad89e869b3/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/1901fa6df621cf9101825636b9ce07df167541bf169b515a170559ad89e869b3/json HTTP/1.1" 200 586 Executing query system flush logs on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query select dictGetFloat64('RedisSimple_flat_Float64_', 'Float64_', toUInt64(1)) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl2 on node1 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query select dictHas('RedisSimple_flat_Float64_', toUInt64(1)) on node Executing query select dictGetDateOrDefault('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '13') ASYNC on node1 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Connection dropped: socket connection error: None Executing query select dictGet('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query SELECT status, error FROM system.backups WHERE id='ab869243-79be-4cb8-8abc-4ffe46fb50fa' on node1 Executing query select dictGetUUID('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:6641 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query rename table system.text_log to system.text_log_31_test on node2 Executing query select dictGetFloat64OrDefault('RedisSimple_flat_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetDateTime('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Connection dropped: socket connection error: None Stdout:6641 Executing query select 20 on instance Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisSimple_flat_Float64_', 'Float64_', toUInt64(2)) on node Executing query rename table system.query_log to system.query_log_31_test on node2 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetFloat64('RedisSimple_flat_Float64_', 'Float64_', toUInt64(2)) on node Executing query system flush logs on node2 Executing query SELECT status, error FROM system.backups WHERE id='ab869243-79be-4cb8-8abc-4ffe46fb50fa' on node1 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictHas('RedisSimple_flat_Float64_', toUInt64(2)) on node Executing query select 20 on instance Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat64OrDefault('RedisSimple_flat_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetDate('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '13') ASYNC on node1 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetString('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on instance Executing query system reload dictionary RedisSimple_hashed_Float64_ on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query rename table system.text_log to system.text_log_32_test on node2 Executing query SELECT status, error FROM system.backups WHERE id='af86eb8e-1877-49ef-9386-3e486630eab7' on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query select dictGet('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(1)) on node Executing query rename table system.query_log to system.query_log_32_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query system flush logs on node2 Executing query DROP DATABASE test_table_backup on instance Executing query select dictGetFloat64('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(1)) on node Executing query select dictGetStringOrDefault('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query DROP DATABASE test_database on instance Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictHas('RedisSimple_hashed_Float64_', toUInt64(1)) on node Executing query SELECT status, error FROM system.backups WHERE id='af86eb8e-1877-49ef-9386-3e486630eab7' on node1 Executing query select dictGet('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 http://172.16.7.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Command:[docker compose --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --project-name roottestdatabasebackup-gw1 --file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml stop --timeout 20] [gw1] PASSED test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] Executing query select dictGetFloat32('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateTime('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetFloat64OrDefault('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl2 on node1 Executing query rename table system.text_log to system.text_log_33_test on node2 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Stopping Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Stopping Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Stopped Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query rename table system.query_log to system.query_log_33_test on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env --project-name roottestaggregationmemoryefficient-gw4 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/docker-compose.yml down --volumes] Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGet('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(2)) on node Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Executing query system flush logs on node2 Executing query select dictGetFloat32OrDefault('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetFloat64('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(2)) on node Executing query SELECT * FROM mydb.tbl2 ORDER BY y on node2 Executing query select dictGet('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictHas('RedisSimple_hashed_Float64_', toUInt64(2)) on node Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Stopping Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Stopping Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Stopped Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Removing Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Stopped Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Removing Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Removed Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Removed Stderr: Network roottestaggregationmemoryefficient-gw4_default Removing Stderr: Network roottestaggregationmemoryefficient-gw4_default Removed Cleanup called Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async Docker networks for project roottestaggregationmemoryefficient-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestaggregationmemoryefficient-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestaggregationmemoryefficient-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestaggregationmemoryefficient-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestaggregationmemoryefficient-gw4 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query select dictGetFloat64('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 test_analyzer_compatibility/test.py::test_two_new_versions Executing query select dictGetOrDefault('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Running tests in /ClickHouse/tests/integration/test_analyzer_compatibility/test.py Cluster start called. is_up=False Docker networks for project roottestanalyzercompatibility-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestanalyzercompatibility-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestanalyzercompatibility-gw4 are DRIVER VOLUME NAME Cleanup called Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Docker networks for project roottestanalyzercompatibility-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestanalyzercompatibility-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestanalyzercompatibility-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestanalyzercompatibility-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestanalyzercompatibility-gw4 Trying to prune unused networks... Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: current Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_analyzer_compatibility/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/configs/config.d Setup database dir /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/database Setup logs dir /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: backward Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_analyzer_compatibility/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/configs/config.d Setup database dir /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/database Setup logs dir /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env --project-name roottestanalyzercompatibility-gw4 --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/docker-compose.yml --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/docker-compose.yml pull] Executing query select dictGetString('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetFloat64OrDefault('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query rename table system.text_log to system.text_log_34_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query rename table system.query_log to system.query_log_34_test on node2 Executing query system reload dictionary RedisSimple_cache_Float64_ on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetFloat64OrDefault('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query system flush logs on node2 Executing query select dictGet('RedisSimple_cache_Float64_', 'Float64_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64('RedisSimple_cache_Float64_', 'Float64_', toUInt64(1)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetUInt8('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisSimple_cache_Float64_', toUInt64(1)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Executing query rename table system.text_log to system.text_log_35_test on node2 Executing query select dictGetFloat32('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetFloat64OrDefault('RedisSimple_cache_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query CREATE TABLE mydb.tbl(x UInt8, y String) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Executing query rename table system.query_log to system.query_log_35_test on node2 Executing query select dictGet('RedisSimple_cache_Float64_', 'Float64_', toUInt64(2)) on node Executing query select dictGetUInt8OrDefault('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query system flush logs on node2 Executing query select dictGetFloat64('RedisSimple_cache_Float64_', 'Float64_', toUInt64(2)) on node Executing query select dictGet('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query SYSTEM SYNC DATABASE REPLICA mydb on node2 Executing query select dictHas('RedisSimple_cache_Float64_', toUInt64(2)) on node Executing query SYSTEM STOP MERGES mydb.tbl on node1 Executing query select dictGetUInt16('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query SYSTEM STOP MERGES mydb.tbl on node2 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SYSTEM STOP FETCHES mydb.tbl on node1 Executing query select dictGetFloat64OrDefault('RedisSimple_cache_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query rename table system.text_log to system.text_log_36_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query select dictGetFloat64('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SYSTEM STOP FETCHES mydb.tbl on node2 Executing query system reload dictionary RedisSimple_direct_Float64_ on node Executing query rename table system.query_log to system.query_log_36_test on node2 Executing query select dictGetUInt16OrDefault('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query INSERT INTO mydb.tbl VALUES (1, 'a') on node1 Executing query select dictGet('RedisSimple_direct_Float64_', 'Float64_', toUInt64(1)) on node Executing query system flush logs on node2 Executing query select dictGet('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictGetFloat64('RedisSimple_direct_Float64_', 'Float64_', toUInt64(1)) on node Executing query INSERT INTO mydb.tbl VALUES (2, 'b') on node1 Executing query select dictGetUInt32('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictHas('RedisSimple_direct_Float64_', toUInt64(1)) on node Executing query INSERT INTO mydb.tbl VALUES (3, 'x') on node2 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] Executing query system reload dictionaries on local_node Executing query select dictGetOrDefault('RedisSimple_direct_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query INSERT INTO mydb.tbl VALUES (4, 'y') on node2 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query select dictGetFloat64OrDefault('RedisSimple_direct_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query rename table system.text_log to system.text_log_37_test on node2 Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Executing query select dictGetUInt32OrDefault('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query select dictGet('RedisSimple_direct_Float64_', 'Float64_', toUInt64(2)) on node Executing query rename table system.query_log to system.query_log_37_test on node2 Executing query select dictGetUInt8('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '14') SETTINGS replica_num=2 on node1 Executing query select dictGet('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64('RedisSimple_direct_Float64_', 'Float64_', toUInt64(2)) on node Executing query system flush logs on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUInt64('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisSimple_direct_Float64_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query select dictGetOrDefault('RedisSimple_direct_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '14') SETTINGS replica_num_in_backup=2 on node1 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetFloat64OrDefault('RedisSimple_direct_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUInt64OrDefault('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictGetUInt16('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query rename table system.text_log to system.text_log_38_test on node2 Executing query system reload dictionary RedisHash_complex_key_hashed_Float64_ on node Executing query rename table system.query_log to system.query_log_38_test on node2 Executing query select dictGet('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 Executing query system flush logs on node2 Executing query select dictGetFloat64('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictGetInt8('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT _part, * FROM mydb.tbl ORDER BY x on node1 Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisHash_complex_key_hashed_Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT _part, * FROM mydb.tbl ORDER BY x on node2 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Executing query select dictGetUInt32('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetInt8OrDefault('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query rename table system.text_log to system.text_log_39_test on node2 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query rename table system.query_log to system.query_log_39_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetFloat64('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt16('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query system flush logs on node2 Executing query select dictHas('RedisHash_complex_key_hashed_Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUInt64('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetInt16OrDefault('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query system reload dictionary RedisHash_complex_key_cache_Float64_ on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query rename table system.text_log to system.text_log_40_test on node2 Executing query select dictGet('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/databases/{database}','{shard}','{replica}') on node1 Executing query rename table system.query_log to system.query_log_40_test on node2 Executing query select dictGetInt32('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node2-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: 2463 ? 00:00:26 clickhouse run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node2-1 bash -c pkill clickhouse] Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('RedisHash_complex_key_cache_Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query CREATE TABLE mydb.tbl(x Int64) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Stdout:2463 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGetInt8('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query INSERT INTO mydb.tbl VALUES (-3) on node1 Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetInt32OrDefault('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query INSERT INTO mydb.tbl VALUES (1) on node1 Executing query select dictGet('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query INSERT INTO mydb.tbl VALUES (10) on node1 run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt64('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Stdout:2463 Executing query select dictGetFloat64('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '15') on node1 Executing query select dictHas('RedisHash_complex_key_cache_Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetInt16('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query RESTORE DATABASE mydb AS mydb2 ON CLUSTER 'cluster' FROM Disk('backups', '15') on node1 Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetInt64OrDefault('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2463 Executing query system reload dictionary RedisHash_complex_key_direct_Float64_ on node Executing query select dictGet('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query select dictGet('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUUID('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query INSERT INTO mydb.tbl VALUES (2) on node1 Executing query select dictGetFloat64('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SYSTEM SYNC DATABASE REPLICA ON CLUSTER 'cluster' mydb2 on node1 Executing query select dictHas('RedisHash_complex_key_direct_Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetInt32('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb2.tbl on node1 Stdout:2463 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetUUIDOrDefault('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Executing query select dictGet('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query select dictGet('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT * FROM mydb2.tbl ORDER BY x on node1 Executing query select dictGetDate('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query select dictGetFloat64('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT * FROM mydb2.tbl ORDER BY x on node2 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisHash_complex_key_direct_Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2463 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictGetInt64('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetDateOrDefault('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGet('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query system reload dictionaries on node Executing query select dictGetDateTime('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] Stdout:2463 Run test with id: 5 Executing query system reload dictionaries on node test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query system reload dictionary RedisSimple_flat_Int16_ on node Executing query select dictGetUUID('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGet('RedisSimple_flat_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/databases/{uuid}','{shard}','{replica}') on node1 Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt16('RedisSimple_flat_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('RedisSimple_flat_Int16_', toUInt64(1)) on node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw9-node2-1/exec HTTP/1.1" 201 74 Executing query CREATE TABLE mydb.tbl(x Int64) ENGINE=ReplicatedMergeTree ORDER BY x on node1 http://localhost:None "POST /v1.46/exec/3e7b4b972ad6997d35cf22b6e7522f536eeabd51df5076f4c4f3459d2d0ad556/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/3e7b4b972ad6997d35cf22b6e7522f536eeabd51df5076f4c4f3459d2d0ad556/json HTTP/1.1" 200 586 Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetString('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisSimple_flat_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node Executing query INSERT INTO mydb.tbl VALUES (-3) on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt16OrDefault('RedisSimple_flat_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query INSERT INTO mydb.tbl VALUES (1) on node1 Executing query select dictGetDate('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGet('RedisSimple_flat_Int16_', 'Int16_', toUInt64(2)) on node Executing query INSERT INTO mydb.tbl VALUES (10) on node1 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetStringOrDefault('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt16('RedisSimple_flat_Int16_', 'Int16_', toUInt64(2)) on node Stdout:3318 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw9-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '16') on node1 Stdout:3318 Executing query select 20 on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGet('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisSimple_flat_Int16_', toUInt64(2)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetFloat32('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node Executing query RESTORE DATABASE mydb AS mydb2 ON CLUSTER 'cluster' FROM Disk('backups', '16') on node1 Executing query select 20 on node2 Executing query select dictGetInt16OrDefault('RedisSimple_flat_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDateTime('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node Executing query system reload dictionary RedisSimple_hashed_Int16_ on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Stderr: backward Pulling Stderr: current Pulling Stderr: 4477f8fe99eb Pulling fs layer Stderr: 84f6d8efd27c Pulling fs layer Stderr: 4f4fb700ef54 Pulling fs layer Stderr: e5c63395e7a3 Pulling fs layer Stderr: 22e20d7710b3 Pulling fs layer Stderr: 184beec54d26 Pulling fs layer Stderr: e5c63395e7a3 Waiting Stderr: 5add26bac827 Pulling fs layer Stderr: 6d305684be88 Pulling fs layer Stderr: 22e20d7710b3 Waiting Stderr: cd532f345afc Pulling fs layer Stderr: 184beec54d26 Waiting Stderr: 5add26bac827 Waiting Stderr: 6d305684be88 Waiting Stderr: cd532f345afc Waiting Stderr: 4f4fb700ef54 Downloading [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Verifying Checksum Stderr: 4f4fb700ef54 Download complete Stderr: current Pulled Stderr: 4477f8fe99eb Downloading [> ] 277.3kB/27.51MB Stderr: 84f6d8efd27c Downloading [> ] 162.9kB/15.37MB Stderr: e5c63395e7a3 Downloading [> ] 533.3kB/276.3MB Stderr: 84f6d8efd27c Verifying Checksum Stderr: 84f6d8efd27c Download complete Stderr: 4477f8fe99eb Downloading [==============================> ] 16.97MB/27.51MB Stderr: 22e20d7710b3 Downloading [=======> ] 3.646kB/24.12kB Stderr: 22e20d7710b3 Downloading [==================================================>] 24.12kB/24.12kB Stderr: 22e20d7710b3 Verifying Checksum Stderr: 22e20d7710b3 Download complete Stderr: e5c63395e7a3 Downloading [==> ] 12.62MB/276.3MB Stderr: 184beec54d26 Downloading [> ] 13.78kB/863.5kB Stderr: 184beec54d26 Verifying Checksum Stderr: 184beec54d26 Download complete Stderr: 5add26bac827 Downloading [==================================================>] 116B/116B Stderr: 5add26bac827 Verifying Checksum Stderr: 5add26bac827 Download complete Stderr: 6d305684be88 Downloading [==================================================>] 362B/362B Stderr: 6d305684be88 Download complete Stderr: cd532f345afc Downloading [==================================================>] 3.09kB/3.09kB Stderr: cd532f345afc Verifying Checksum Stderr: cd532f345afc Download complete Stderr: 4477f8fe99eb Verifying Checksum Stderr: 4477f8fe99eb Download complete Stderr: e5c63395e7a3 Downloading [=====> ] 29.88MB/276.3MB Stderr: 4477f8fe99eb Extracting [> ] 294.9kB/27.51MB Stderr: e5c63395e7a3 Downloading [========> ] 48.8MB/276.3MB Stderr: 4477f8fe99eb Extracting [=======> ] 4.129MB/27.51MB Stderr: e5c63395e7a3 Downloading [============> ] 66.6MB/276.3MB Stderr: 4477f8fe99eb Extracting [==============> ] 7.963MB/27.51MB Stderr: e5c63395e7a3 Downloading [==============> ] 79.95MB/276.3MB Stderr: 4477f8fe99eb Extracting [==================> ] 10.03MB/27.51MB Stderr: e5c63395e7a3 Downloading [==================> ] 101.1MB/276.3MB Stderr: 4477f8fe99eb Extracting [===========================> ] 15.04MB/27.51MB Stderr: e5c63395e7a3 Downloading [=====================> ] 119.5MB/276.3MB Stderr: 4477f8fe99eb Extracting [===================================> ] 19.76MB/27.51MB Stderr: e5c63395e7a3 Downloading [========================> ] 136.7MB/276.3MB Stderr: 4477f8fe99eb Extracting [=========================================> ] 23MB/27.51MB Stderr: e5c63395e7a3 Downloading [============================> ] 156.6MB/276.3MB Stderr: e5c63395e7a3 Downloading [==============================> ] 170MB/276.3MB Stderr: 4477f8fe99eb Extracting [==========================================> ] 23.3MB/27.51MB Stderr: e5c63395e7a3 Downloading [==================================> ] 188.9MB/276.3MB Stderr: 4477f8fe99eb Extracting [==============================================> ] 25.66MB/27.51MB Stderr: e5c63395e7a3 Downloading [=====================================> ] 205.6MB/276.3MB Stderr: 4477f8fe99eb Extracting [================================================> ] 26.54MB/27.51MB Stderr: e5c63395e7a3 Downloading [=========================================> ] 226.8MB/276.3MB Stderr: 4477f8fe99eb Extracting [=================================================> ] 27.13MB/27.51MB Stderr: e5c63395e7a3 Downloading [============================================> ] 244.1MB/276.3MB Stderr: 4477f8fe99eb Extracting [==================================================>] 27.51MB/27.51MB Stderr: e5c63395e7a3 Downloading [===============================================> ] 264.7MB/276.3MB Stderr: 4477f8fe99eb Pull complete Stderr: 84f6d8efd27c Extracting [> ] 163.8kB/15.37MB Stderr: e5c63395e7a3 Downloading [==================================================>] 276.3MB/276.3MB Stderr: e5c63395e7a3 Verifying Checksum Stderr: e5c63395e7a3 Download complete Stderr: 84f6d8efd27c Extracting [======> ] 1.966MB/15.37MB Stderr: 84f6d8efd27c Extracting [============> ] 3.932MB/15.37MB Stderr: 84f6d8efd27c Extracting [=============================> ] 9.175MB/15.37MB Stderr: 84f6d8efd27c Extracting [====================================> ] 11.3MB/15.37MB Stderr: 84f6d8efd27c Extracting [==========================================> ] 12.94MB/15.37MB Stderr: 84f6d8efd27c Extracting [==============================================> ] 14.25MB/15.37MB Stderr: 84f6d8efd27c Extracting [===============================================> ] 14.75MB/15.37MB Stderr: 84f6d8efd27c Extracting [================================================> ] 14.91MB/15.37MB Stderr: 84f6d8efd27c Extracting [=================================================> ] 15.07MB/15.37MB Stderr: 84f6d8efd27c Extracting [==================================================>] 15.37MB/15.37MB Stderr: 84f6d8efd27c Pull complete Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Pull complete Stderr: e5c63395e7a3 Extracting [> ] 557.1kB/276.3MB Stderr: e5c63395e7a3 Extracting [=> ] 7.799MB/276.3MB Stderr: e5c63395e7a3 Extracting [==> ] 15.6MB/276.3MB Stderr: e5c63395e7a3 Extracting [====> ] 22.28MB/276.3MB Stderr: e5c63395e7a3 Extracting [=====> ] 28.97MB/276.3MB Stderr: e5c63395e7a3 Extracting [======> ] 33.42MB/276.3MB Stderr: e5c63395e7a3 Extracting [=======> ] 38.99MB/276.3MB Stderr: e5c63395e7a3 Extracting [=======> ] 43.45MB/276.3MB Stderr: e5c63395e7a3 Extracting [========> ] 47.91MB/276.3MB Stderr: e5c63395e7a3 Extracting [=========> ] 52.92MB/276.3MB Stderr: e5c63395e7a3 Extracting [==========> ] 57.93MB/276.3MB Stderr: e5c63395e7a3 Extracting [===========> ] 63.5MB/276.3MB Stderr: e5c63395e7a3 Extracting [============> ] 68.52MB/276.3MB Stderr: e5c63395e7a3 Extracting [=============> ] 73.53MB/276.3MB Stderr: e5c63395e7a3 Extracting [==============> ] 78.54MB/276.3MB Stderr: e5c63395e7a3 Extracting [===============> ] 84.12MB/276.3MB Stderr: e5c63395e7a3 Extracting [================> ] 90.24MB/276.3MB Stderr: e5c63395e7a3 Extracting [=================> ] 98.04MB/276.3MB Stderr: e5c63395e7a3 Extracting [===================> ] 107.5MB/276.3MB Stderr: e5c63395e7a3 Extracting [====================> ] 115.9MB/276.3MB Stderr: e5c63395e7a3 Extracting [======================> ] 123.7MB/276.3MB Stderr: e5c63395e7a3 Extracting [=======================> ] 131.5MB/276.3MB Stderr: e5c63395e7a3 Extracting [========================> ] 134.8MB/276.3MB Stderr: e5c63395e7a3 Extracting [========================> ] 135.9MB/276.3MB Stderr: e5c63395e7a3 Extracting [========================> ] 137.6MB/276.3MB Stderr: e5c63395e7a3 Extracting [=========================> ] 139.3MB/276.3MB Stderr: e5c63395e7a3 Extracting [==========================> ] 148.7MB/276.3MB Stderr: e5c63395e7a3 Extracting [============================> ] 157.6MB/276.3MB Stderr: e5c63395e7a3 Extracting [==============================> ] 166MB/276.3MB Stderr: e5c63395e7a3 Extracting [===============================> ] 173.2MB/276.3MB Stderr: e5c63395e7a3 Extracting [================================> ] 180.5MB/276.3MB Stderr: e5c63395e7a3 Extracting [=================================> ] 187.2MB/276.3MB Stderr: e5c63395e7a3 Extracting [===================================> ] 195.5MB/276.3MB Stderr: e5c63395e7a3 Extracting [====================================> ] 203.3MB/276.3MB Stderr: e5c63395e7a3 Extracting [======================================> ] 210.6MB/276.3MB Stderr: e5c63395e7a3 Extracting [======================================> ] 212.8MB/276.3MB Stderr: e5c63395e7a3 Extracting [=======================================> ] 219.5MB/276.3MB Stderr: e5c63395e7a3 Extracting [=========================================> ] 227.3MB/276.3MB Stderr: e5c63395e7a3 Extracting [==========================================> ] 232.8MB/276.3MB Stderr: e5c63395e7a3 Extracting [===========================================> ] 237.9MB/276.3MB Stderr: e5c63395e7a3 Extracting [============================================> ] 244.5MB/276.3MB Stderr: e5c63395e7a3 Extracting [=============================================> ] 251.2MB/276.3MB Stderr: e5c63395e7a3 Extracting [==============================================> ] 256.2MB/276.3MB Stderr: e5c63395e7a3 Extracting [===============================================> ] 262.4MB/276.3MB Stderr: e5c63395e7a3 Extracting [================================================> ] 267.9MB/276.3MB Stderr: e5c63395e7a3 Extracting [=================================================> ] 271.3MB/276.3MB Stderr: e5c63395e7a3 Extracting [=================================================> ] 275.2MB/276.3MB Stderr: e5c63395e7a3 Extracting [==================================================>] 276.3MB/276.3MB Stderr: e5c63395e7a3 Pull complete Stderr: 22e20d7710b3 Extracting [==================================================>] 24.12kB/24.12kB Stderr: 22e20d7710b3 Extracting [==================================================>] 24.12kB/24.12kB Stderr: 22e20d7710b3 Pull complete Stderr: 184beec54d26 Extracting [=> ] 32.77kB/863.5kB Stderr: 184beec54d26 Extracting [==================================================>] 863.5kB/863.5kB Stderr: 184beec54d26 Extracting [==================================================>] 863.5kB/863.5kB Stderr: 184beec54d26 Pull complete Stderr: 5add26bac827 Extracting [==================================================>] 116B/116B Stderr: 5add26bac827 Extracting [==================================================>] 116B/116B Stderr: 5add26bac827 Pull complete Stderr: 6d305684be88 Extracting [==================================================>] 362B/362B Stderr: 6d305684be88 Extracting [==================================================>] 362B/362B Stderr: 6d305684be88 Pull complete Stderr: cd532f345afc Extracting [==================================================>] 3.09kB/3.09kB Stderr: cd532f345afc Extracting [==================================================>] 3.09kB/3.09kB Stderr: cd532f345afc Pull complete Stderr: backward Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env --project-name roottestanalyzercompatibility-gw4 --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/docker-compose.yml --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env --project-name roottestanalyzercompatibility-gw4 --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/docker-compose.yml --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/docker-compose.yml up -d --no-recreate] Executing query select dictGet('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(1)) on node Executing query select 20 on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Stderr: Container roottestdatabasebackup-gw1-instance-1 Stopping Stderr: Container roottestdatabasebackup-gw1-resolver-1 Stopping Stderr: Container roottestdatabasebackup-gw1-instance-1 Stopped Stderr: Container roottestdatabasebackup-gw1-minio1-1 Stopping Stderr: Container roottestdatabasebackup-gw1-minio1-1 Stopped Stderr: Container roottestdatabasebackup-gw1-resolver-1 Stopped Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Stopping Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Stopping Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Stopped Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --project-name roottestdatabasebackup-gw1 --file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml down --volumes] Executing query select dictGetFloat32OrDefault('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query INSERT INTO mydb.tbl VALUES (2) on node1 Executing query select dictGetInt16('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGet('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SYSTEM SYNC DATABASE REPLICA ON CLUSTER 'cluster' mydb2 on node1 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisSimple_hashed_Int16_', toUInt64(1)) on node Executing query select dictGetFloat64('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select 20 on node2 Stderr: Container roottestdatabasebackup-gw1-resolver-1 Stopping Stderr: Container roottestdatabasebackup-gw1-instance-1 Stopping Stderr: Container roottestdatabasebackup-gw1-resolver-1 Stopped Stderr: Container roottestdatabasebackup-gw1-resolver-1 Removing Stderr: Container roottestdatabasebackup-gw1-instance-1 Stopped Stderr: Container roottestdatabasebackup-gw1-instance-1 Removing Stderr: Container roottestdatabasebackup-gw1-instance-1 Removed Stderr: Container roottestdatabasebackup-gw1-minio1-1 Stopping Stderr: Container roottestdatabasebackup-gw1-minio1-1 Stopped Stderr: Container roottestdatabasebackup-gw1-minio1-1 Removing Stderr: Container roottestdatabasebackup-gw1-resolver-1 Removed Stderr: Container roottestdatabasebackup-gw1-minio1-1 Removed Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Stopping Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Stopping Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Stopped Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Removing Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Stopped Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Removing Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Removed Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Removed Stderr: Volume roottestdatabasebackup-gw1_data1-1 Removing Stderr: Network roottestdatabasebackup-gw1_default Removing Stderr: Volume roottestdatabasebackup-gw1_data1-1 Removed Stderr: Network roottestdatabasebackup-gw1_default Removed Cleanup called Stderr: Network roottestanalyzercompatibility-gw4_default Creating Stderr: Network roottestanalyzercompatibility-gw4_default Created Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Creating Stderr: Container roottestanalyzercompatibility-gw4-current-1 Creating Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Created Stderr: Container roottestanalyzercompatibility-gw4-current-1 Created Stderr: Container roottestanalyzercompatibility-gw4-current-1 Starting Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Starting Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Started Stderr: Container roottestanalyzercompatibility-gw4-current-1 Started ClickHouse instance created get_instance_ip instance_name=current http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-current-1/json HTTP/1.1" 200 None get_instance_ip instance_name=current http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-current-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in current, ip: 172.16.2.3... http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-current-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query select dictGetString('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Docker networks for project roottestdatabasebackup-gw1 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestdatabasebackup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestdatabasebackup-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdatabasebackup-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestdatabasebackup-gw1 Trying to prune unused networks... http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb2.tbl on node1 Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Stdout:3 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 3 http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query drop table if exists system.text_log_1_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetInt16OrDefault('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Executing query drop table if exists system.query_log_1_test on node2 http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query select dictGet('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(2)) on node http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query select dictGetFloat64OrDefault('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query SELECT * FROM mydb2.tbl ORDER BY x on node1 Executing query drop table if exists system.text_log_2_test on node2 http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query select dictGetInt16('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(2)) on node http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query drop table if exists system.query_log_2_test on node2 http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] Executing query system reload dictionaries on node test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] Executing query SELECT * FROM mydb2.tbl ORDER BY x on node2 http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisSimple_hashed_Int16_', toUInt64(2)) on node http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query drop table if exists system.text_log_3_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetFloat32('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query drop table if exists system.query_log_3_test on node2 http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query select dictGetUInt8('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "GET /v1.46/containers/55100ef433e266028138d98cdd20913f4c557d73e795a4d402cc768250b72bee/json HTTP/1.1" 200 None ClickHouse current started get_instance_ip instance_name=backward http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-backward-1/json HTTP/1.1" 200 None get_instance_ip instance_name=backward http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-backward-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in backward, ip: 172.16.2.2... http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-backward-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c05b433d7bdf4a2eda8b46407fd1d8cce3397842c2d19f4bb15f17080da7db9c/json HTTP/1.1" 200 None ClickHouse backward started Executing query select dictGetInt16OrDefault('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query drop table if exists system.text_log_4_test on node2 Executing query SYSTEM FLUSH LOGS on current Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.query_log_4_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query system reload dictionary RedisSimple_cache_Int16_ on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query drop table if exists system.text_log_5_test on node2 Executing query select dictGet('RedisSimple_cache_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query select dictGetUInt8OrDefault('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query drop table if exists system.query_log_5_test on node2 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetInt16('RedisSimple_cache_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query drop table if exists system.text_log_6_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Executing query SYSTEM FLUSH LOGS on backward Executing query select dictHas('RedisSimple_cache_Int16_', toUInt64(1)) on node Executing query drop table if exists system.query_log_6_test on node2 Executing query select dictGetFloat64('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetUInt16('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisSimple_cache_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query SELECT name FROM clusterAllReplicas('test_cluster_mixed', system.tables); on current Executing query drop table if exists system.text_log_7_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16OrDefault('RedisSimple_cache_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node test_backup_restore_on_cluster/test.py::test_replicated_table Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query SYSTEM FLUSH LOGS on current Executing query drop table if exists system.query_log_7_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictGet('RedisSimple_cache_Int16_', 'Int16_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query drop table if exists system.text_log_8_test on node2 Executing query INSERT INTO tbl VALUES (1, 'Don''t') on node1 Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictGetInt16('RedisSimple_cache_Int16_', 'Int16_', toUInt64(2)) on node Executing query select dictGetUInt16OrDefault('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query drop table if exists system.query_log_8_test on node2 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query INSERT INTO tbl VALUES (2, 'count') on node2 Executing query select dictHas('RedisSimple_cache_Int16_', toUInt64(2)) on node Executing query SYSTEM FLUSH LOGS on backward Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.text_log_9_test on node2 Executing query INSERT INTO tbl SETTINGS async_insert=true VALUES (3, 'your') on node1 Executing query select dictGetUInt8('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query drop table if exists system.query_log_9_test on node2 Executing query select dictGetUInt32('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT hostname() AS h, getSetting('allow_experimental_analyzer') FROM clusterAllReplicas('test_cluster_mixed', system.one) ORDER BY h; on current Executing query INSERT INTO tbl SETTINGS async_insert=true VALUES (4, 'chickens') on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt16OrDefault('RedisSimple_cache_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query drop table if exists system.text_log_10_test on node2 Executing query SELECT DISTINCT Settings['allow_experimental_analyzer'] FROM system.query_log WHERE initial_query_id = '47a6d5d3-1fe6-4501-aacd-d07315653d1b'; on backward Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.query_log_10_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query system reload dictionary RedisSimple_direct_Int16_ on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query SELECT name FROM clusterAllReplicas('test_cluster_mixed', system.tables) on backward Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query drop table if exists system.text_log_11_test on node2 Executing query select dictGet('RedisSimple_direct_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query SYSTEM FLUSH LOGS on current Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '17') SETTINGS replica_num=1 on node1 Executing query select dictGetUInt32OrDefault('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query drop table if exists system.query_log_11_test on node2 Executing query select dictGetInt16('RedisSimple_direct_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.text_log_12_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisSimple_direct_Int16_', toUInt64(1)) on node Executing query select dictGetUInt16('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query drop table if exists system.query_log_12_test on node2 Executing query select dictGetUInt64('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_direct_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query SYSTEM FLUSH LOGS on backward Executing query drop table if exists system.text_log_13_test on node2 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '17') on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictGetInt16OrDefault('RedisSimple_direct_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query SELECT hostname() AS h, getSetting('allow_experimental_analyzer') FROM clusterAllReplicas('test_cluster_mixed', system.one) ORDER BY h; on backward Executing query drop table if exists system.query_log_13_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictGet('RedisSimple_direct_Int16_', 'Int16_', toUInt64(2)) on node Executing query SELECT DISTINCT Settings['allow_experimental_analyzer'] FROM clusterAllReplicas('test_cluster_mixed', system.query_log) WHERE initial_query_id = '95d58f6d-fec6-40d2-a892-0867586f7e1a'; on backward Executing query drop table if exists system.text_log_14_test on node2 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetUInt64OrDefault('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictGetInt16('RedisSimple_direct_Int16_', 'Int16_', toUInt64(2)) on node Executing query SELECT name FROM clusterAllReplicas('test_cluster_mixed', system.tables) SETTINGS enable_analyzer = 1; on current Executing query drop table if exists system.query_log_14_test on node2 Executing query select dictGetUInt32('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisSimple_direct_Int16_', toUInt64(2)) on node Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query drop table if exists system.text_log_15_test on node2 Executing query SYSTEM FLUSH LOGS on current Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt8('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisSimple_direct_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query drop table if exists system.query_log_15_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16OrDefault('RedisSimple_direct_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query drop table if exists system.text_log_16_test on node2 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query SYSTEM FLUSH LOGS on backward Executing query system reload dictionary RedisHash_complex_key_hashed_Int16_ on node Executing query drop table if exists system.query_log_16_test on node2 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetInt8OrDefault('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGet('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT DISTINCT Settings['allow_experimental_analyzer'] FROM system.query_log WHERE initial_query_id = '6b3c0f3f-ecb2-44c2-9f92-919919e4aede'; on current Executing query drop table if exists system.text_log_17_test on node2 Executing query select dictGetUInt64('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query drop table if exists system.query_log_17_test on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env --project-name roottestanalyzercompatibility-gw4 --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/docker-compose.yml --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/docker-compose.yml stop --timeout 20] [gw4] PASSED test_analyzer_compatibility/test.py::test_two_new_versions Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisHash_complex_key_hashed_Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.text_log_18_test on node2 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.query_log_18_test on node2 Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetInt16OrDefault('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.text_log_19_test on node2 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt16OrDefault('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.query_log_19_test on node2 test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x Int32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictGetInt8('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.text_log_20_test on node2 Executing query select dictGetInt16('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query INSERT INTO tbl VALUES (1) on node1 Executing query select dictGetInt32('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.query_log_20_test on node2 Executing query select dictHas('RedisHash_complex_key_hashed_Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query ALTER TABLE tbl ADD COLUMN y Int32 on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.text_log_21_test on node2 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query INSERT INTO tbl VALUES (2, 20) on node1 Executing query drop table if exists system.query_log_21_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGetInt16OrDefault('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query ALTER TABLE tbl ADD COLUMN z Int32 on node1 Executing query drop table if exists system.text_log_22_test on node2 Executing query select dictGetInt32OrDefault('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query system reload dictionary RedisHash_complex_key_cache_Int16_ on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.query_log_22_test on node2 Executing query INSERT INTO tbl VALUES (3, 30, 300) on node1 Executing query select dictGet('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.text_log_23_test on node2 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '18') on node1 Connection dropped: socket connection error: None Executing query select dictGetInt16('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt64('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.query_log_23_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisHash_complex_key_cache_Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.text_log_24_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.query_log_24_test on node2 Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query drop table if exists system.text_log_25_test on node2 Executing query select dictGetInt64OrDefault('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetInt16OrDefault('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '18') on node1 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.query_log_25_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt32('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.text_log_26_test on node2 Executing query select dictGetUUID('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query OPTIMIZE TABLE tbl FINAL on node1 Executing query drop table if exists system.query_log_26_test on node2 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisHash_complex_key_cache_Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query drop table if exists system.text_log_27_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query drop table if exists system.query_log_27_test on node2 Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query select dictGetUUIDOrDefault('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetInt16OrDefault('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.text_log_28_test on node2 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters Executing query select dictGet('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.query_log_28_test on node2 Executing query system reload dictionary RedisHash_complex_key_direct_Int16_ on node Executing query drop table if exists system.text_log_29_test on node2 Executing query select dictGetDate('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt64('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query drop table if exists system.query_log_29_test on node2 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.text_log_30_test on node2 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictHas('RedisHash_complex_key_direct_Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query drop table if exists system.query_log_30_test on node2 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetDateOrDefault('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.text_log_31_test on node2 Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16OrDefault('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.query_log_31_test on node2 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDateTime('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.text_log_32_test on node2 test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictGetUUID('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.query_log_32_test on node2 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.text_log_33_test on node2 Executing query INSERT INTO tbl VALUES (111) on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('RedisHash_complex_key_direct_Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Stopping Stderr: Container roottestanalyzercompatibility-gw4-current-1 Stopping Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Stopped Stderr: Container roottestanalyzercompatibility-gw4-current-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env --project-name roottestanalyzercompatibility-gw4 --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/docker-compose.yml --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/docker-compose.yml down --volumes] Executing query drop table if exists system.query_log_33_test on node2 Executing query INSERT INTO tbl VALUES (222) on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.text_log_34_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '19') on node1 Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetInt16OrDefault('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.query_log_34_test on node2 Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Stopping Stderr: Container roottestanalyzercompatibility-gw4-current-1 Stopping Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Stopped Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Removing Stderr: Container roottestanalyzercompatibility-gw4-current-1 Stopped Stderr: Container roottestanalyzercompatibility-gw4-current-1 Removing Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Removed Stderr: Container roottestanalyzercompatibility-gw4-current-1 Removed Stderr: Network roottestanalyzercompatibility-gw4_default Removing Stderr: Network roottestanalyzercompatibility-gw4_default Removed Cleanup called Executing query select dictGetString('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Docker networks for project roottestanalyzercompatibility-gw4 are NETWORK ID NAME DRIVER SCOPE Executing query system reload dictionaries on node Docker containers for project roottestanalyzercompatibility-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestanalyzercompatibility-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestanalyzercompatibility-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestanalyzercompatibility-gw4 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:3 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 3 Executing query drop table if exists system.text_log_35_test on node2 Executing query select dictGetDate('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node [gw2] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] Run test with id: 6 Executing query system reload dictionaries on node Executing query drop table if exists system.query_log_35_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query system reload dictionary RedisSimple_flat_Int32_ on node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster3' FROM Disk('backups', '19') on node1 Executing query drop table if exists system.text_log_36_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGetStringOrDefault('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGet('RedisSimple_flat_Int32_', 'Int32_', toUInt64(1)) on node Executing query drop table if exists system.query_log_36_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Connection dropped: socket connection error: None Executing query select dictGetInt32('RedisSimple_flat_Int32_', 'Int32_', toUInt64(1)) on node Executing query drop table if exists system.text_log_37_test on node2 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster3' tbl on node1 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetFloat32('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisSimple_flat_Int32_', toUInt64(1)) on node Executing query drop table if exists system.query_log_37_test on node2 Executing query select dictGetDateTime('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetOrDefault('RedisSimple_flat_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query drop table if exists system.text_log_38_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt32OrDefault('RedisSimple_flat_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query drop table if exists system.query_log_38_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGet('RedisSimple_flat_Int32_', 'Int32_', toUInt64(2)) on node Executing query drop table if exists system.text_log_39_test on node2 Executing query SELECT * FROM tbl ORDER BY x on node3 Executing query select dictGetFloat32OrDefault('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetInt32('RedisSimple_flat_Int32_', 'Int32_', toUInt64(2)) on node Executing query drop table if exists system.query_log_39_test on node2 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster Executing query select dictGet('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisSimple_flat_Int32_', toUInt64(2)) on node Executing query drop table if exists system.text_log_40_test on node2 Executing query select dictGetFloat64('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('RedisSimple_flat_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query drop table if exists system.query_log_40_test on node2 Executing query select dictGetString('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt32OrDefault('RedisSimple_flat_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node [gw9] PASSED test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables Executing query create database lazy engine=Lazy(10) on node1 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query create database a on node1 Executing query system reload dictionary RedisSimple_hashed_Int32_ on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query create table lazy.src (n int, m int) engine=Log on node1 Executing query select dictGet('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(1)) on node Executing query select dictGetFloat64OrDefault('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query create dictionary a.d (n int default 0, m int default 42) primary key n source(clickhouse(host 'localhost' port tcpPort() user 'default' table 'src' password '' db 'lazy'))lifetime(min 1 max 10) layout(flat()) on node1 Executing query select dictGetInt32('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(1)) on node Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query create table system.join (n int, m int) engine=Join(any, left, n) on node1 test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictHas('RedisSimple_hashed_Int32_', toUInt64(1)) on node Executing query select dictGetUInt8('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat32('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query insert into system.join values (1, 1) on node1 Executing query select dictGetOrDefault('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query INSERT INTO tbl VALUES (111) on node1 Executing query insert into system.join values (1, 2) on node1 Executing query select dictGetInt32OrDefault('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query INSERT INTO tbl VALUES (222) on node2 Executing query insert into system.join values (1, 3) on node1 Executing query select dictGet('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(2)) on node Executing query select dictGetUInt8OrDefault('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '20') on node1 Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query insert into system.join values (1, 4) on node1 Executing query select dictGetInt32('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(2)) on node Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query insert into system.join values (1, 5) on node1 Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetUInt16('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisSimple_hashed_Int32_', toUInt64(2)) on node Executing query select dictGetFloat64('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query insert into system.join values (1, 6) on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster1' FROM Disk('backups', '20') on node1 Executing query insert into system.join values (1, 7) on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query select dictGetInt32OrDefault('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query insert into system.join values (1, 8) on node1 Executing query select dictGetUInt16OrDefault('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query system reload dictionary RedisSimple_cache_Int32_ on node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query insert into system.join values (1, 9) on node1 Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster Executing query select dictGet('RedisSimple_cache_Int32_', 'Int32_', toUInt64(1)) on node [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] Executing query system reload dictionaries on local_node Executing query insert into system.join values (1, 10) on node1 Executing query select dictGetUInt32('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt32('RedisSimple_cache_Int32_', 'Int32_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'KeyField2', toUInt64(1), toDate('2019-02-10')) on local_node Executing query insert into system.join values (1, 11) on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisSimple_cache_Int32_', toUInt64(1)) on node Executing query select dictGetDate('LocalClickHouse_range_hashed_', 'KeyField2', toUInt64(1), toDate('2019-02-10')) on local_node Executing query insert into system.join values (1, 12) on node1 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query select dictGetOrDefault('RedisSimple_cache_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt8_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query insert into system.join values (1, 13) on node1 Executing query select dictGetUInt32OrDefault('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query select dictGetInt32OrDefault('RedisSimple_cache_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query insert into system.join values (1, 14) on node1 Executing query select dictGetUInt8('LocalClickHouse_range_hashed_', 'UInt8_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('RedisSimple_cache_Int32_', 'Int32_', toUInt64(2)) on node Executing query insert into system.join values (1, 15) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt16_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetInt32('RedisSimple_cache_Int32_', 'Int32_', toUInt64(2)) on node Executing query select dictGetUInt64('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 16) on node1 Executing query select dictGetUInt16('LocalClickHouse_range_hashed_', 'UInt16_', toUInt64(1), toDate('2019-02-10')) on local_node test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def Executing query CREATE TABLE tbl (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY tuple() on node1 Executing query select dictHas('RedisSimple_cache_Int32_', toUInt64(2)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 17) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt32_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query CREATE TABLE tbl (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY tuple() on node2 Executing query select dictGetOrDefault('RedisSimple_cache_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query insert into system.join values (1, 18) on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictGetUInt32('LocalClickHouse_range_hashed_', 'UInt32_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query SYSTEM STOP REPLICATION QUEUES tbl on node2 Executing query insert into system.join values (1, 19) on node1 Executing query select dictGetInt32OrDefault('RedisSimple_cache_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetUInt64OrDefault('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt64_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query ALTER TABLE tbl MODIFY COLUMN x String on node1 Executing query insert into system.join values (1, 20) on node1 Executing query system reload dictionary RedisSimple_direct_Int32_ on node Executing query select dictGet('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt64('LocalClickHouse_range_hashed_', 'UInt64_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query insert into system.join values (1, 21) on node1 Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node1 Executing query select dictGet('RedisSimple_direct_Int32_', 'Int32_', toUInt64(1)) on node Executing query select dictGetInt8('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int8_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query insert into system.join values (1, 22) on node1 Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node2 Executing query select dictGetInt32('RedisSimple_direct_Int32_', 'Int32_', toUInt64(1)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt8('LocalClickHouse_range_hashed_', 'Int8_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query insert into system.join values (1, 23) on node1 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '21') on node2 Executing query select dictHas('RedisSimple_direct_Int32_', toUInt64(1)) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int16_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query insert into system.join values (1, 24) on node1 Executing query select dictGetOrDefault('RedisSimple_direct_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetInt8OrDefault('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetInt16('LocalClickHouse_range_hashed_', 'Int16_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query insert into system.join values (1, 25) on node1 Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetInt32OrDefault('RedisSimple_direct_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int32_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 26) on node1 Executing query select dictGet('RedisSimple_direct_Int32_', 'Int32_', toUInt64(2)) on node Executing query select dictGetInt16('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 27) on node1 Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '21') SETTINGS replica_num_in_backup=1 on node1 Executing query select dictGetInt32('LocalClickHouse_range_hashed_', 'Int32_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetInt32('RedisSimple_direct_Int32_', 'Int32_', toUInt64(2)) on node Executing query insert into system.join values (1, 28) on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int64_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictHas('RedisSimple_direct_Int32_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query insert into system.join values (1, 29) on node1 Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node1 Executing query select dictGetInt64('LocalClickHouse_range_hashed_', 'Int64_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetOrDefault('RedisSimple_direct_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node2 Executing query select dictGet('LocalClickHouse_range_hashed_', 'UUID_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetInt16OrDefault('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query insert into system.join values (1, 30) on node1 Executing query select dictGetInt32OrDefault('RedisSimple_direct_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 31) on node1 Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetUUID('LocalClickHouse_range_hashed_', 'UUID_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query system reload dictionary RedisHash_complex_key_hashed_Int32_ on node Executing query select dictGetInt32('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 32) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Date_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '21') SETTINGS replica_num_in_backup=2 on node2 Executing query select dictGet('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDate('LocalClickHouse_range_hashed_', 'Date_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query insert into system.join values (1, 33) on node1 Executing query select dictGetInt32('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query insert into system.join values (1, 34) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'DateTime_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node1 Executing query select dictGetInt32OrDefault('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictHas('RedisHash_complex_key_hashed_Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query insert into system.join values (1, 35) on node1 Executing query select dictGetDateTime('LocalClickHouse_range_hashed_', 'DateTime_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query insert into system.join values (1, 36) on node1 Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node2 Executing query select dictGet('LocalClickHouse_range_hashed_', 'String_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetInt64('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt32OrDefault('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query insert into system.join values (1, 37) on node1 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def Executing query select dictGetString('LocalClickHouse_range_hashed_', 'String_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 38) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Float32_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetInt32('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 39) on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetFloat32('LocalClickHouse_range_hashed_', 'Float32_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetInt64OrDefault('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictHas('RedisHash_complex_key_hashed_Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 40) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Float64_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGet('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 41) on node1 Executing query select dictGetFloat64('LocalClickHouse_range_hashed_', 'Float64_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetInt32OrDefault('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGetUUID('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 42) on node1 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query system reload dictionary RedisHash_complex_key_cache_Int32_ on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'KeyField2', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 43) on node1 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetDate('LocalClickHouse_range_hashed_', 'KeyField2', toUInt64(2), toDate('2019-04-10')) on local_node Executing query insert into system.join values (1, 44) on node1 Executing query select dictGetInt32('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt8_', toUInt64(2), toDate('2019-04-10')) on local_node test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictGetUUIDOrDefault('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query insert into system.join values (1, 45) on node1 Executing query select dictHas('RedisHash_complex_key_cache_Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt8('LocalClickHouse_range_hashed_', 'UInt8_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 46) on node1 Executing query INSERT INTO tbl VALUES (111) on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt16_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetDate('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 47) on node1 Executing query select dictGetInt32OrDefault('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query INSERT INTO tbl VALUES (222) on node2 Executing query select dictGetUInt16('LocalClickHouse_range_hashed_', 'UInt16_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query insert into system.join values (1, 48) on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 49) on node1 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictGetInt32('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 50) on node1 Executing query select dictGetUInt32('LocalClickHouse_range_hashed_', 'UInt32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query SYSTEM STOP REPLICATED SENDS ON CLUSTER 'cluster' tbl on node1 Executing query select dictGetDateOrDefault('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictHas('RedisHash_complex_key_cache_Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 51) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query INSERT INTO tbl VALUES (333) on node1 Executing query select dictGetUInt64('LocalClickHouse_range_hashed_', 'UInt64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query insert into system.join values (1, 52) on node1 Executing query select dictGetDateTime('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt32OrDefault('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query INSERT INTO tbl VALUES (444) on node2 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int8_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query insert into system.join values (1, 53) on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '22') on node1 Executing query system reload dictionary RedisHash_complex_key_direct_Int32_ on node Executing query insert into system.join values (1, 54) on node1 Executing query select dictGetInt8('LocalClickHouse_range_hashed_', 'Int8_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGet('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query insert into system.join values (1, 55) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int16_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt32('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query insert into system.join values (1, 56) on node1 Executing query select dictGetInt16('LocalClickHouse_range_hashed_', 'Int16_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisHash_complex_key_direct_Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query insert into system.join values (1, 57) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetString('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '22') on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query insert into system.join values (1, 58) on node1 Executing query select dictGetInt32('LocalClickHouse_range_hashed_', 'Int32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt32OrDefault('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query insert into system.join values (1, 59) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictGet('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 60) on node1 Executing query select dictGetInt64('LocalClickHouse_range_hashed_', 'Int64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetStringOrDefault('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGetInt32('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 61) on node1 Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'UUID_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 62) on node1 Executing query select dictHas('RedisHash_complex_key_direct_Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query select dictGetUUID('LocalClickHouse_range_hashed_', 'UUID_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetFloat32('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 63) on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert Executing query select dictGet('LocalClickHouse_range_hashed_', 'Date_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 64) on node1 Executing query select dictGetInt32OrDefault('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query insert into system.join values (1, 65) on node1 Executing query select dictGetDate('LocalClickHouse_range_hashed_', 'Date_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query system reload dictionaries on node Executing query select dictGetFloat32OrDefault('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query insert into system.join values (1, 66) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'DateTime_', toUInt64(2), toDate('2019-04-10')) on local_node [gw2] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] Run test with id: 7 Executing query system reload dictionaries on node Executing query select dictGet('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 67) on node1 Executing query select dictGetDateTime('LocalClickHouse_range_hashed_', 'DateTime_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query system reload dictionary RedisSimple_flat_Int64_ on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query insert into system.join values (1, 68) on node1 Executing query select dictGetFloat64('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'String_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('RedisSimple_flat_Int64_', 'Int64_', toUInt64(1)) on node Executing query insert into system.join values (1, 69) on node1 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetString('LocalClickHouse_range_hashed_', 'String_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetInt64('RedisSimple_flat_Int64_', 'Int64_', toUInt64(1)) on node Executing query insert into system.join values (1, 70) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Float32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictHas('RedisSimple_flat_Int64_', toUInt64(1)) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query insert into system.join values (1, 71) on node1 Executing query select dictGetFloat32('LocalClickHouse_range_hashed_', 'Float32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetFloat64OrDefault('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query insert into system.join values (1, 72) on node1 Executing query SYSTEM STOP MERGES ON CLUSTER 'cluster' tbl on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Float64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetInt64OrDefault('RedisSimple_flat_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] Executing query system reload dictionaries on node test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] Executing query insert into system.join values (1, 73) on node1 Executing query select dictGetFloat64('LocalClickHouse_range_hashed_', 'Float64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query INSERT INTO tbl VALUES (111) on node1 Executing query select dictGet('RedisSimple_flat_Int64_', 'Int64_', toUInt64(2)) on node Executing query select dictGet('MySQL_range_hashed_', 'KeyField2', toUInt64(1), toDate('2019-02-10')) on node Executing query insert into system.join values (1, 74) on node1 [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] Executing query system reload dictionaries on local_node Executing query INSERT INTO tbl VALUES (222) on node1 Executing query select dictGetInt64('RedisSimple_flat_Int64_', 'Int64_', toUInt64(2)) on node Executing query insert into system.join values (1, 75) on node1 Executing query select dictGetDate('MySQL_range_hashed_', 'KeyField2', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGet('LocalClickHouse_cache_', 'UInt8_', toUInt64(1)) on local_node Executing query SYSTEM SYNC REPLICA tbl on node2 Executing query select dictHas('RedisSimple_flat_Int64_', toUInt64(2)) on node Executing query insert into system.join values (1, 76) on node1 Executing query select dictGet('MySQL_range_hashed_', 'UInt8_', toUInt64(1), toDate('2019-02-10')) on node Executing query SYSTEM START MERGES tbl on node2 Executing query select dictGetOrDefault('RedisSimple_flat_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetUInt8('LocalClickHouse_cache_', 'UInt8_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 77) on node1 Executing query OPTIMIZE TABLE tbl FINAL on node2 Executing query select dictGetInt64OrDefault('RedisSimple_flat_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetUInt8('MySQL_range_hashed_', 'UInt8_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 78) on node1 Executing query system reload dictionary RedisSimple_hashed_Int64_ on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '23') on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'UInt16_', toUInt64(1), toDate('2019-02-10')) on node Executing query insert into system.join values (1, 79) on node1 Executing query select dictGet('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(1)) on node Executing query select dictGetUInt16('MySQL_range_hashed_', 'UInt16_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query insert into system.join values (1, 80) on node1 Executing query select dictGetInt64('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(1)) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGet('MySQL_range_hashed_', 'UInt32_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGet('LocalClickHouse_cache_', 'UInt16_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 81) on node1 Executing query select dictHas('RedisSimple_hashed_Int64_', toUInt64(1)) on node Executing query select dictGetUInt32('MySQL_range_hashed_', 'UInt32_', toUInt64(1), toDate('2019-02-10')) on node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '23') on node1 Executing query select dictGetUInt16('LocalClickHouse_cache_', 'UInt16_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 82) on node1 Executing query select dictGetOrDefault('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGet('MySQL_range_hashed_', 'UInt64_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 83) on node1 Executing query select dictGetInt64OrDefault('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetUInt64('MySQL_range_hashed_', 'UInt64_', toUInt64(1), toDate('2019-02-10')) on node Executing query insert into system.join values (1, 84) on node1 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictGet('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(2)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Int8_', toUInt64(1), toDate('2019-02-10')) on node Executing query insert into system.join values (1, 85) on node1 Executing query select dictGetInt64('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(2)) on node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetInt8('MySQL_range_hashed_', 'Int8_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGet('LocalClickHouse_cache_', 'UInt32_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 86) on node1 Executing query select dictHas('RedisSimple_hashed_Int64_', toUInt64(2)) on node Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query select dictGet('MySQL_range_hashed_', 'Int16_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetUInt32('LocalClickHouse_cache_', 'UInt32_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 87) on node1 Executing query select dictGetOrDefault('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge Executing query select dictGetInt16('MySQL_range_hashed_', 'Int16_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 88) on node1 Executing query select dictGetInt64OrDefault('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGet('MySQL_range_hashed_', 'Int32_', toUInt64(1), toDate('2019-02-10')) on node Executing query insert into system.join values (1, 89) on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query system reload dictionary RedisSimple_cache_Int64_ on node Executing query select dictGetInt32('MySQL_range_hashed_', 'Int32_', toUInt64(1), toDate('2019-02-10')) on node Executing query insert into system.join values (1, 90) on node1 Executing query select dictGet('RedisSimple_cache_Int64_', 'Int64_', toUInt64(1)) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('MySQL_range_hashed_', 'Int64_', toUInt64(1), toDate('2019-02-10')) on node Executing query insert into system.join values (1, 91) on node1 Executing query select dictGetInt64('RedisSimple_cache_Int64_', 'Int64_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_cache_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictGetInt64('MySQL_range_hashed_', 'Int64_', toUInt64(1), toDate('2019-02-10')) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('RedisSimple_cache_Int64_', toUInt64(1)) on node Executing query select dictGetUInt64('LocalClickHouse_cache_', 'UInt64_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 92) on node1 Executing query select dictGet('MySQL_range_hashed_', 'UUID_', toUInt64(1), toDate('2019-02-10')) on node Executing query insert into system.join values (1, 93) on node1 Executing query select dictGetOrDefault('RedisSimple_cache_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetUUID('MySQL_range_hashed_', 'UUID_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetInt64OrDefault('RedisSimple_cache_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query insert into system.join values (1, 94) on node1 test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/{uuid}','{replica}')ORDER BY x on node1 Executing query select dictGet('MySQL_range_hashed_', 'Date_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGet('RedisSimple_cache_Int64_', 'Int64_', toUInt64(2)) on node Executing query insert into system.join values (1, 95) on node1 Executing query select dictGetUInt64OrDefault('LocalClickHouse_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetDate('MySQL_range_hashed_', 'Date_', toUInt64(1), toDate('2019-02-10')) on node Executing query INSERT INTO tbl VALUES (1, 'AA') on node1 Executing query select dictGetInt64('RedisSimple_cache_Int64_', 'Int64_', toUInt64(2)) on node Executing query insert into system.join values (1, 96) on node1 Executing query select dictGet('LocalClickHouse_cache_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'DateTime_', toUInt64(1), toDate('2019-02-10')) on node Executing query INSERT INTO tbl VALUES (2, 'BB') on node2 Executing query insert into system.join values (1, 97) on node1 Executing query select dictHas('RedisSimple_cache_Int64_', toUInt64(2)) on node Executing query select dictGetInt8('LocalClickHouse_cache_', 'Int8_', toUInt64(1)) on local_node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '24') on node1 Executing query select dictGetDateTime('MySQL_range_hashed_', 'DateTime_', toUInt64(1), toDate('2019-02-10')) on node Executing query insert into system.join values (1, 98) on node1 Executing query select dictGetOrDefault('RedisSimple_cache_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'String_', toUInt64(1), toDate('2019-02-10')) on node Executing query insert into system.join values (1, 99) on node1 Executing query select dictGetInt64OrDefault('RedisSimple_cache_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetString('MySQL_range_hashed_', 'String_', toUInt64(1), toDate('2019-02-10')) on node Executing query RESTORE TABLE tbl AS tbl2 ON CLUSTER 'cluster' FROM Disk('backups', '24') on node2 Executing query create table src (n int, m default joinGet('system.join', 'm', 1::int),t default dictGetOrNull('a.d', 'm', toUInt64(3)),k default dictGet('a.d', 'm', toUInt64(4))) engine=MergeTree order by n on node1 Executing query system reload dictionary RedisSimple_direct_Int64_ on node Executing query select dictGetInt8OrDefault('LocalClickHouse_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Float32_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGet('RedisSimple_direct_Int64_', 'Int64_', toUInt64(1)) on node Executing query create dictionary test.d (n int default 0, m int default 42) primary key n source(clickhouse(host 'localhost' port tcpPort() user 'default' table 'src' password '' db 'default'))lifetime(min 1 max 10) layout(flat()) on node1 Connection dropped: socket connection error: None Executing query select dictGet('LocalClickHouse_cache_', 'Int16_', toUInt64(1)) on local_node Executing query select dictGetFloat32('MySQL_range_hashed_', 'Float32_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetInt64('RedisSimple_direct_Int64_', 'Int64_', toUInt64(1)) on node Executing query select dictGetInt16('LocalClickHouse_cache_', 'Int16_', toUInt64(1)) on local_node Executing query INSERT INTO tbl2 VALUES (3, 'CC') on node1 Executing query create table join (n int, m default dictGet('a.d', 'm', toUInt64(3)),k default dictGet('test.d', 'm', toUInt64(0))) engine=Join(any, left, n) on node1 Executing query select dictGet('MySQL_range_hashed_', 'Float64_', toUInt64(1), toDate('2019-02-10')) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictHas('RedisSimple_direct_Int64_', toUInt64(1)) on node Executing query create table lazy.log (n default dictGet(test.d, 'm', toUInt64(0))) engine=Log on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGetFloat64('MySQL_range_hashed_', 'Float64_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetOrDefault('RedisSimple_direct_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl2 on node1 Executing query create table a.t (n default joinGet('system.join', 'm', 1::int),m default dictGet('test.d', 'm', toUInt64(3)),k default joinGet(join, 'm', 1::int)) engine=MergeTree order by n on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'KeyField2', toUInt64(2), toDate('2019-04-10')) on node Executing query select database || '.' || name from system.tables on node1 Executing query select dictGetInt64OrDefault('RedisSimple_direct_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetDate('MySQL_range_hashed_', 'KeyField2', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGet('RedisSimple_direct_Int64_', 'Int64_', toUInt64(2)) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node1-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:28 clickhouse run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node1-1 bash -c pkill clickhouse] Executing query select dictGet('LocalClickHouse_cache_', 'Int32_', toUInt64(1)) on local_node Executing query SELECT * FROM tbl2 ORDER BY x on node1 run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGet('MySQL_range_hashed_', 'UInt8_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetInt64('RedisSimple_direct_Int64_', 'Int64_', toUInt64(2)) on node Executing query select dictGetInt32('LocalClickHouse_cache_', 'Int32_', toUInt64(1)) on local_node Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query select dictGetUInt8('MySQL_range_hashed_', 'UInt8_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictHas('RedisSimple_direct_Int64_', toUInt64(2)) on node Executing query SELECT * FROM tbl2 ORDER BY x on node2 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'UInt16_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetOrDefault('RedisSimple_direct_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetInt64OrDefault('RedisSimple_direct_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetUInt16('MySQL_range_hashed_', 'UInt16_', toUInt64(2), toDate('2019-04-10')) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetInt32OrDefault('LocalClickHouse_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query system reload dictionary RedisHash_complex_key_hashed_Int64_ on node Executing query select dictGet('MySQL_range_hashed_', 'UInt32_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int64_', toUInt64(1)) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUInt32('MySQL_range_hashed_', 'UInt32_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetInt64('LocalClickHouse_cache_', 'Int64_', toUInt64(1)) on local_node Executing query select dictGetInt64('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('MySQL_range_hashed_', 'UInt64_', toUInt64(2), toDate('2019-04-10')) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictHas('RedisHash_complex_key_hashed_Int64_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetUInt64('MySQL_range_hashed_', 'UInt64_', toUInt64(2), toDate('2019-04-10')) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Int8_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetInt64OrDefault('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node test_backup_restore_on_cluster/test.py::test_required_privileges Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictGetInt8('MySQL_range_hashed_', 'Int8_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGet('LocalClickHouse_cache_', 'UUID_', toUInt64(1)) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('MySQL_range_hashed_', 'Int16_', toUInt64(2), toDate('2019-04-10')) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query INSERT INTO tbl VALUES (100) on node1 Executing query select dictGetUUID('LocalClickHouse_cache_', 'UUID_', toUInt64(1)) on local_node Stdout:8 Executing query select dictGetInt64('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt16('MySQL_range_hashed_', 'Int16_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query CREATE USER u1 on node1 Executing query select dictHas('RedisHash_complex_key_hashed_Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('MySQL_range_hashed_', 'Int32_', toUInt64(2), toDate('2019-04-10')) on node Executing query GRANT CLUSTER ON *.* TO u1 on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetInt32('MySQL_range_hashed_', 'Int32_', toUInt64(2), toDate('2019-04-10')) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '25') on node1 Executing query select dictGetInt64OrDefault('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Int64_', toUInt64(2), toDate('2019-04-10')) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw9-node1-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/b877b6095aa28494a2196b990f078567455269c7e55dd1b86b13fee14b7d4626/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/b877b6095aa28494a2196b990f078567455269c7e55dd1b86b13fee14b7d4626/json HTTP/1.1" 200 586 Executing query system reload dictionary RedisHash_complex_key_cache_Int64_ on node Executing query select dictGet('LocalClickHouse_cache_', 'Date_', toUInt64(1)) on local_node Executing query GRANT BACKUP ON tbl TO u1 on node1 Executing query select dictGetInt64('MySQL_range_hashed_', 'Int64_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGet('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDate('LocalClickHouse_cache_', 'Date_', toUInt64(1)) on local_node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '25') on node1 Executing query select dictGet('MySQL_range_hashed_', 'UUID_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetInt64('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGetUUID('MySQL_range_hashed_', 'UUID_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictHas('RedisHash_complex_key_cache_Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGet('MySQL_range_hashed_', 'Date_', toUInt64(2), toDate('2019-04-10')) on node Stdout:766 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:766 Executing query select 20 on node1 Executing query select dictGetDateOrDefault('LocalClickHouse_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetInt64OrDefault('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetDate('MySQL_range_hashed_', 'Date_', toUInt64(2), toDate('2019-04-10')) on node Executing query RESTORE TABLE tbl AS tbl2 ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 Executing query select dictGet('LocalClickHouse_cache_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictGet('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('MySQL_range_hashed_', 'DateTime_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetDateTime('LocalClickHouse_cache_', 'DateTime_', toUInt64(1)) on local_node Executing query GRANT INSERT, CREATE TABLE ON tbl2 TO u1 on node1 Executing query select 20 on node1 Executing query select dictGetInt64('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDateTime('MySQL_range_hashed_', 'DateTime_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query RESTORE TABLE tbl AS tbl2 ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 Executing query select dictHas('RedisHash_complex_key_cache_Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('MySQL_range_hashed_', 'String_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetString('MySQL_range_hashed_', 'String_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select 20 on node1 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl2 on node2 Executing query select dictGetInt64OrDefault('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGet('MySQL_range_hashed_', 'Float32_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGet('LocalClickHouse_cache_', 'String_', toUInt64(1)) on local_node Executing query select database || '.' || name from system.tables on node1 Executing query system reload dictionary RedisHash_complex_key_direct_Int64_ on node Executing query SELECT * FROM tbl2 on node2 Executing query select dictGetFloat32('MySQL_range_hashed_', 'Float32_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetString('LocalClickHouse_cache_', 'String_', toUInt64(1)) on local_node Executing query DROP TABLE tbl2 ON CLUSTER 'cluster' SYNC on node1 Executing query drop table a.t on node1 Executing query select dictGet('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('MySQL_range_hashed_', 'Float64_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query drop table lazy.log on node1 Executing query REVOKE ALL FROM u1 on node1 Executing query select dictGetInt64('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat64('MySQL_range_hashed_', 'Float64_', toUInt64(2), toDate('2019-04-10')) on node Executing query drop table join on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query RESTORE ALL ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] Executing query system reload dictionaries on node Executing query select dictHas('RedisHash_complex_key_direct_Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query drop dictionary test.d on node1 Executing query select dictGetStringOrDefault('LocalClickHouse_cache_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGet('MySQL_cache_', 'UInt8_', toUInt64(1)) on node Executing query GRANT INSERT, CREATE TABLE ON tbl TO u1 on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query drop table src on node1 Executing query select dictGet('LocalClickHouse_cache_', 'Float32_', toUInt64(1)) on local_node Executing query RESTORE ALL ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 Executing query select dictGetUInt8('MySQL_cache_', 'UInt8_', toUInt64(1)) on node Executing query drop table system.join on node1 Executing query select dictGetInt64OrDefault('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetFloat32('LocalClickHouse_cache_', 'Float32_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query drop database a on node1 Executing query select dictGet('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node2 Executing query select dictGetOrDefault('MySQL_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query drop database lazy on node1 Executing query select dictGetInt64('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query SELECT * FROM tbl on node2 [gw9] PASSED test_async_load_databases/test.py::test_dependent_tables Executing query CREATE TABLE test.elements (id UInt64, a String, b Int32, c Float64) ENGINE=Log; on node1 test_async_load_databases/test.py::test_dict_get_data Executing query select dictGetUInt8OrDefault('MySQL_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictHas('RedisHash_complex_key_direct_Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query INSERT INTO test.elements VALUES (0, 'water', 10, 1), (1, 'air', 40, 0.01), (2, 'earth', 100, 1.7); on node1 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_required_privileges Executing query select dictGet('MySQL_cache_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Float64_', toUInt64(1)) on local_node Executing query SELECT status FROM system.dictionaries WHERE name='dep_x' on node1 Executing query select dictGetUInt16('MySQL_cache_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetInt64OrDefault('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetFloat64('LocalClickHouse_cache_', 'Float64_', toUInt64(1)) on local_node Executing query SELECT status FROM system.dictionaries WHERE name='dep_y' on node1 Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query system reload dictionaries on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query SELECT status FROM system.dictionaries WHERE name='dep_z' on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node [gw2] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] Run test with id: 4 Executing query system reload dictionaries on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query SELECT dictGetString('dep_x', 'a', toUInt64(1)) on node1 Executing query select dictGetUInt16OrDefault('MySQL_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query system reload dictionary RedisSimple_flat_Int8_ on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGet('MySQL_cache_', 'UInt32_', toUInt64(1)) on node Executing query SELECT status FROM system.dictionaries WHERE name='dep_x' on node1 Executing query select dictGet('RedisSimple_flat_Int8_', 'Int8_', toUInt64(1)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select dictGetUInt32('MySQL_cache_', 'UInt32_', toUInt64(1)) on node Executing query SELECT status FROM system.dictionaries WHERE name='dep_y' on node1 Executing query select dictGetInt8('RedisSimple_flat_Int8_', 'Int8_', toUInt64(1)) on node Executing query select dictGetUInt64('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(1)) on local_node Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query SELECT status FROM system.dictionaries WHERE name='dep_z' on node1 Executing query select dictHas('RedisSimple_flat_Int8_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query SELECT dictGetString('dep_y', 'a', toUInt64(1)) on node1 Executing query INSERT INTO tbl VALUES (3) on node1 Executing query select dictGetOrDefault('RedisSimple_flat_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetUInt32OrDefault('MySQL_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query SELECT dictGetString('dep_z', 'a', toUInt64(1)) on node1 Executing query INSERT INTO tbl VALUES (5) on node2 Executing query select dictGetInt8OrDefault('RedisSimple_flat_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGet('MySQL_cache_', 'UInt64_', toUInt64(1)) on node Executing query SELECT dictGetString('dep_x', 'a', toUInt64(3)) on node1 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '26') ASYNC on node1 Executing query select dictGet('RedisSimple_flat_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_cache_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictGetInt8('RedisSimple_flat_Int8_', 'Int8_', toUInt64(2)) on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps -C clickhouse] Executing query select dictGetUInt64('MySQL_cache_', 'UInt64_', toUInt64(1)) on node Executing query SELECT dictGetString('dep_y', 'a', toUInt64(3)) on node1 Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:46 clickhouse run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestbackuprestoreoncluster-gw0-node2-1 bash -c pkill clickhouse] run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetUInt8('LocalClickHouse_cache_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictHas('RedisSimple_flat_Int8_', toUInt64(2)) on node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query SELECT dictGetString('dep_z', 'a', toUInt64(3)) on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetOrDefault('MySQL_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query INSERT INTO test.elements VALUES (3, 'fire', 30, 8) on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetInt8OrDefault('RedisSimple_flat_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetUInt64OrDefault('MySQL_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictGetUInt8OrDefault('LocalClickHouse_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query system reload dictionary RedisSimple_hashed_Int8_ on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('MySQL_cache_', 'Int8_', toUInt64(1)) on node Stdout:8 Executing query select dictGet('LocalClickHouse_cache_', 'UInt16_', toUInt64(2)) on local_node Executing query select dictGet('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(1)) on node Executing query select dictGetInt8('MySQL_cache_', 'Int8_', toUInt64(1)) on node Executing query select dictGetUInt16('LocalClickHouse_cache_', 'UInt16_', toUInt64(2)) on local_node Executing query select dictGetInt8('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(1)) on node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query select dictHas('RedisSimple_hashed_Int8_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetOrDefault('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Stdout:8 Executing query select dictGetInt8OrDefault('MySQL_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetInt8OrDefault('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGet('MySQL_cache_', 'Int16_', toUInt64(1)) on node Executing query select dictGet('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_cache_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictGetInt8('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGetInt16('MySQL_cache_', 'Int16_', toUInt64(1)) on node Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictGetUInt32('LocalClickHouse_cache_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictHas('RedisSimple_hashed_Int8_', toUInt64(2)) on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Stdout:8 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetOrDefault('MySQL_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGetInt8OrDefault('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetInt16OrDefault('MySQL_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query system reload dictionary RedisSimple_cache_Int8_ on node Executing query select dictGet('MySQL_cache_', 'Int32_', toUInt64(1)) on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('RedisSimple_cache_Int8_', 'Int8_', toUInt64(1)) on node Stdout:8 Executing query select dictGet('LocalClickHouse_cache_', 'UInt64_', toUInt64(2)) on local_node Executing query select dictGetInt32('MySQL_cache_', 'Int32_', toUInt64(1)) on node Executing query select dictGetInt8('RedisSimple_cache_Int8_', 'Int8_', toUInt64(1)) on node Executing query select dictGetUInt64('LocalClickHouse_cache_', 'UInt64_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query select dictHas('RedisSimple_cache_Int8_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetOrDefault('RedisSimple_cache_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt32OrDefault('MySQL_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetInt8OrDefault('RedisSimple_cache_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Stdout:8 Executing query select dictGetUInt64OrDefault('LocalClickHouse_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGet('MySQL_cache_', 'Int64_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int8_', toUInt64(2)) on local_node Executing query select dictGet('RedisSimple_cache_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGetInt64('MySQL_cache_', 'Int64_', toUInt64(1)) on node Executing query select dictGetInt8('LocalClickHouse_cache_', 'Int8_', toUInt64(2)) on local_node Executing query select dictGetInt8('RedisSimple_cache_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query select dictHas('RedisSimple_cache_Int8_', toUInt64(2)) on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetOrDefault('RedisSimple_cache_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetInt64OrDefault('MySQL_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetInt8OrDefault('RedisSimple_cache_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query system reload dictionary RedisSimple_direct_Int8_ on node Executing query select dictGet('MySQL_cache_', 'UUID_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int16_', toUInt64(2)) on local_node Executing query select dictGet('RedisSimple_direct_Int8_', 'Int8_', toUInt64(1)) on node Executing query select dictGetUUID('MySQL_cache_', 'UUID_', toUInt64(1)) on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetInt16('LocalClickHouse_cache_', 'Int16_', toUInt64(2)) on local_node Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictGetInt8('RedisSimple_direct_Int8_', 'Int8_', toUInt64(1)) on node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Connection dropped: socket connection error: None Executing query select dictHas('RedisSimple_direct_Int8_', toUInt64(1)) on node Executing query select dictGetOrDefault('MySQL_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetOrDefault('RedisSimple_direct_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetUUIDOrDefault('MySQL_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt8OrDefault('RedisSimple_direct_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Stdout:8 Executing query select dictGet('MySQL_cache_', 'Date_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGet('RedisSimple_direct_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGetDate('MySQL_cache_', 'Date_', toUInt64(1)) on node Executing query select dictGetInt32('LocalClickHouse_cache_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGetInt8('RedisSimple_direct_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictHas('RedisSimple_direct_Int8_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetOrDefault('RedisSimple_direct_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetDateOrDefault('MySQL_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetInt8OrDefault('RedisSimple_direct_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictGet('MySQL_cache_', 'DateTime_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int64_', toUInt64(2)) on local_node Executing query system reload dictionary RedisHash_complex_key_hashed_Int8_ on node Executing query select dictGetDateTime('MySQL_cache_', 'DateTime_', toUInt64(1)) on node Executing query select dictGetInt64('LocalClickHouse_cache_', 'Int64_', toUInt64(2)) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query SELECT status FROM system.backups WHERE id='45f1880a-b40f-4728-9795-48686a580b15' AND status == 'CREATING_BACKUP' on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetInt8('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('MySQL_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query SELECT status FROM system.backups WHERE id='45f1880a-b40f-4728-9795-48686a580b15' on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictHas('RedisHash_complex_key_hashed_Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateTimeOrDefault('MySQL_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackuprestoreoncluster-gw0-node2-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/95f6c6abcddb4343d2a09e0fda2c0a055aef0e1d1c1cd36ce85f4a22c8fd1f34/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/95f6c6abcddb4343d2a09e0fda2c0a055aef0e1d1c1cd36ce85f4a22c8fd1f34/json HTTP/1.1" 200 586 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGet('MySQL_cache_', 'String_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_cache_', 'UUID_', toUInt64(2)) on local_node Executing query select dictGetInt8OrDefault('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetString('MySQL_cache_', 'String_', toUInt64(1)) on node Executing query select dictGetUUID('LocalClickHouse_cache_', 'UUID_', toUInt64(2)) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetInt8('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:881 Clickhouse process running. run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Stdout:881 Executing query select 20 on node2 Executing query select dictGetStringOrDefault('MySQL_cache_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictHas('RedisHash_complex_key_hashed_Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGet('MySQL_cache_', 'Float32_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Date_', toUInt64(2)) on local_node Executing query select dictGetFloat32('MySQL_cache_', 'Float32_', toUInt64(1)) on node Executing query select dictGetInt8OrDefault('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select 20 on node2 Executing query select dictGetDate('LocalClickHouse_cache_', 'Date_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query system reload dictionary RedisHash_complex_key_cache_Int8_ on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetFloat32OrDefault('MySQL_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetInt8('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on node2 Executing query select dictGet('MySQL_cache_', 'Float64_', toUInt64(1)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictHas('RedisHash_complex_key_cache_Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_cache_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictGetFloat64('MySQL_cache_', 'Float64_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select 20 on node2 Executing query select dictGetDateTime('LocalClickHouse_cache_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query SELECT dictGetString('dep_x', 'a', toUInt64(3)) on node1 Executing query select dictGetInt8OrDefault('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query SELECT dictGetString('dep_y', 'a', toUInt64(3)) on node1 Executing query select dictGet('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetInt8('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT dictGetString('dep_z', 'a', toUInt64(3)) on node1 Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '26') on node1 Executing query select dictGetFloat64OrDefault('MySQL_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictHas('RedisHash_complex_key_cache_Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('MySQL_cache_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query INSERT INTO test.elements VALUES (4, 'ether', 404, 0.001) on node1 Executing query select dictGetUInt64('MySQL_cache_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_cache_', 'String_', toUInt64(2)) on local_node Executing query SELECT dictHas('dep_y', toUInt64(4)) on node1 Executing query SYSTEM SYNC REPLICA tbl on node1 Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query select dictGetString('LocalClickHouse_cache_', 'String_', toUInt64(2)) on local_node Executing query select dictGetInt8OrDefault('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query system reload dictionary RedisHash_complex_key_direct_Int8_ on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup Executing query select dictGet('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt64OrDefault('MySQL_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('MySQL_cache_', 'UInt8_', toUInt64(2)) on node Executing query select dictGetInt8('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetStringOrDefault('LocalClickHouse_cache_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetUInt8('MySQL_cache_', 'UInt8_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Float32_', toUInt64(2)) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('RedisHash_complex_key_direct_Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetFloat32('LocalClickHouse_cache_', 'Float32_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetInt8OrDefault('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query SELECT dictHas('dep_y', toUInt64(4)) on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetUInt8OrDefault('MySQL_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGet('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node test_backup_restore_on_cluster/test.py::test_system_functions Executing query CREATE FUNCTION linear_equation AS (x, k, b) -> k*x + b; on node1 Executing query select dictGet('MySQL_cache_', 'UInt16_', toUInt64(2)) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetInt8('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query CREATE FUNCTION parity_str AS (n) -> if(n % 2, 'odd', 'even'); on node1 Executing query select dictGet('LocalClickHouse_cache_', 'Float64_', toUInt64(2)) on local_node Executing query select dictGetUInt16('MySQL_cache_', 'UInt16_', toUInt64(2)) on node Executing query select dictHas('RedisHash_complex_key_direct_Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query BACKUP TABLE system.functions ON CLUSTER 'cluster' TO Disk('backups', '27') on node1 Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetFloat64('LocalClickHouse_cache_', 'Float64_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetInt8OrDefault('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query DROP FUNCTION linear_equation on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query DROP FUNCTION parity_str on node1 Executing query system reload dictionaries on node Executing query select dictGet('MySQL_cache_', 'UInt32_', toUInt64(2)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query SELECT name FROM system.functions WHERE name='parity_str' on node2 [gw2] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] test_dictionaries_redis/test.py::test_redis_dictionaries[String_] Run test with id: 11 Executing query system reload dictionaries on node Executing query select dictGetUInt32('MySQL_cache_', 'UInt32_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(2)) on local_node Executing query RESTORE TABLE system.functions ON CLUSTER 'cluster' FROM Disk('backups', '27') on node1 Executing query SELECT dictHas('dep_y', toUInt64(4)) on node1 Executing query system reload dictionary RedisSimple_flat_String_ on node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetUInt64('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGet('RedisSimple_flat_String_', 'String_', toUInt64(1)) on node Executing query SELECT dictGetString('dep_x', 'a', toUInt64(4)) on node1 Executing query SELECT number, linear_equation(number, 2, 1) FROM numbers(3) on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetString('RedisSimple_flat_String_', 'String_', toUInt64(1)) on node Executing query SELECT dictGetString('dep_y', 'a', toUInt64(4)) on node1 Executing query SELECT number, parity_str(number) FROM numbers(3) on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetUInt32OrDefault('MySQL_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictHas('RedisSimple_flat_String_', toUInt64(1)) on node Executing query SELECT dictGetString('dep_z', 'a', toUInt64(4)) on node1 Executing query SELECT number, linear_equation(number, 2, 1) FROM numbers(3) on node2 Executing query select dictGetUInt64OrDefault('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGet('MySQL_cache_', 'UInt64_', toUInt64(2)) on node Executing query select dictGetOrDefault('RedisSimple_flat_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query DROP TABLE IF EXISTS test.elements; on node1 Executing query SELECT number, parity_str(number) FROM numbers(3) on node2 Executing query select dictGetHierarchy('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGetUInt64('MySQL_cache_', 'UInt64_', toUInt64(2)) on node Executing query select dictGetStringOrDefault('RedisSimple_flat_String_', 'String_', toUInt64(9999999), toString('hi')) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node1-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: 766 ? 00:00:14 clickhouse run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node1-1 bash -c pkill clickhouse] Executing query SELECT name FROM system.functions WHERE name='parity_str' on node2 run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:766 Executing query select dictGetHierarchy('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGet('RedisSimple_flat_String_', 'String_', toUInt64(2)) on node Executing query SELECT number, parity_str(number) FROM numbers(3) on node2 Executing query select dictIsIn('LocalClickHouse_cache_', toUInt64(1), toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetString('RedisSimple_flat_String_', 'String_', toUInt64(2)) on node Executing query DROP FUNCTION linear_equation on node1 Executing query select dictIsIn('LocalClickHouse_cache_', toUInt64(2), toUInt64(1)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictHas('RedisSimple_flat_String_', toUInt64(2)) on node Executing query DROP FUNCTION parity_str on node1 Executing query select dictGet('MySQL_cache_', 'Int8_', toUInt64(2)) on node [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] Executing query system reload dictionaries on local_node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('RedisSimple_flat_String_', 'String_', toUInt64(9999999), toString('hi')) on node Stdout:766 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_system_functions Executing query select dictGetInt8('MySQL_cache_', 'Int8_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictGetStringOrDefault('RedisSimple_flat_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUInt8('LocalClickHouse_direct_', 'UInt8_', toUInt64(1)) on local_node Executing query system reload dictionary RedisSimple_hashed_String_ on node Executing query select dictGetOrDefault('MySQL_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGet('RedisSimple_hashed_String_', 'String_', toUInt64(1)) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt8OrDefault('MySQL_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Stdout:766 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetString('RedisSimple_hashed_String_', 'String_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGet('MySQL_cache_', 'Int16_', toUInt64(2)) on node Executing query select dictHas('RedisSimple_hashed_String_', toUInt64(1)) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetInt16('MySQL_cache_', 'Int16_', toUInt64(2)) on node Executing query select dictGetOrDefault('RedisSimple_hashed_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'UInt16_', toUInt64(1)) on local_node test_backup_restore_on_cluster/test.py::test_system_users Executing query CREATE USER u1 SETTINGS custom_a=123 on node1 Executing query select dictGetStringOrDefault('RedisSimple_hashed_String_', 'String_', toUInt64(9999999), toString('hi')) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('MySQL_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Stdout:766 Executing query GRANT SELECT ON tbl TO u1 on node1 Executing query select dictGetUInt16('LocalClickHouse_direct_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictGet('RedisSimple_hashed_String_', 'String_', toUInt64(2)) on node Executing query select dictGetInt16OrDefault('MySQL_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query CREATE USER u2 SETTINGS allow_backup=false on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGetString('RedisSimple_hashed_String_', 'String_', toUInt64(2)) on node Executing query select dictGet('MySQL_cache_', 'Int32_', toUInt64(2)) on node Executing query GRANT CLUSTER ON *.* TO u2 on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictHas('RedisSimple_hashed_String_', toUInt64(2)) on node Executing query BACKUP TABLE system.users ON CLUSTER 'cluster' TO Disk('backups', '28') on node1 Executing query select dictGetInt32('MySQL_cache_', 'Int32_', toUInt64(2)) on node Executing query select dictGetOrDefault('RedisSimple_hashed_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:766 Stdout:1540 Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query GRANT BACKUP ON system.users TO u2 on node1 Executing query select dictGet('LocalClickHouse_direct_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictGetStringOrDefault('RedisSimple_hashed_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetOrDefault('MySQL_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query BACKUP TABLE system.users ON CLUSTER 'cluster' TO Disk('backups', '28') on node1 Executing query system reload dictionary RedisSimple_cache_String_ on node Executing query select dictGetUInt32('LocalClickHouse_direct_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictGetInt32OrDefault('MySQL_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('RedisSimple_cache_String_', 'String_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query DROP USER u1 on node1 Executing query select dictGet('MySQL_cache_', 'Int64_', toUInt64(2)) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetString('RedisSimple_cache_String_', 'String_', toUInt64(1)) on node Executing query RESTORE TABLE system.users ON CLUSTER 'cluster' FROM Disk('backups', '28') on node1 Executing query select dictGetInt64('MySQL_cache_', 'Int64_', toUInt64(2)) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw9-node1-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/a6d367f337081ef4007f9990f3f3b8442e7551516b7726bf50ffe13f4319bfcc/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/a6d367f337081ef4007f9990f3f3b8442e7551516b7726bf50ffe13f4319bfcc/json HTTP/1.1" 200 586 Executing query select dictHas('RedisSimple_cache_String_', toUInt64(1)) on node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query GRANT CREATE USER ON *.* TO u2 on node1 Executing query select dictGetOrDefault('RedisSimple_cache_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetOrDefault('MySQL_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGet('LocalClickHouse_direct_', 'UInt64_', toUInt64(1)) on local_node Executing query RESTORE TABLE system.users ON CLUSTER 'cluster' FROM Disk('backups', '28') on node1 Executing query select dictGetInt64OrDefault('MySQL_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetStringOrDefault('RedisSimple_cache_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetUInt64('LocalClickHouse_direct_', 'UInt64_', toUInt64(1)) on local_node Executing query GRANT SELECT ON tbl TO u2 WITH GRANT OPTION on node1 run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('MySQL_cache_', 'UUID_', toUInt64(2)) on node Executing query select dictGet('RedisSimple_cache_String_', 'String_', toUInt64(2)) on node Stdout:1577 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1577 Executing query select 20 on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query RESTORE TABLE system.users ON CLUSTER 'cluster' FROM Disk('backups', '28') on node1 Executing query select dictGetUUID('MySQL_cache_', 'UUID_', toUInt64(2)) on node Executing query select dictGetString('RedisSimple_cache_String_', 'String_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictHas('RedisSimple_cache_String_', toUInt64(2)) on node Executing query SHOW CREATE USER u1 on node1 Executing query select dictGetUInt64OrDefault('LocalClickHouse_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetOrDefault('RedisSimple_cache_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select 20 on node1 Executing query SHOW GRANTS FOR u1 on node1 Executing query select dictGet('LocalClickHouse_direct_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGetUUIDOrDefault('MySQL_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetStringOrDefault('RedisSimple_cache_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_system_users Executing query system reload dictionary RedisSimple_direct_String_ on node Executing query select dictGetInt8('LocalClickHouse_direct_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_cache_', 'Date_', toUInt64(2)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('RedisSimple_direct_String_', 'String_', toUInt64(1)) on node Executing query select 20 on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGetDate('MySQL_cache_', 'Date_', toUInt64(2)) on node Executing query select dictGetString('RedisSimple_direct_String_', 'String_', toUInt64(1)) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictHas('RedisSimple_direct_String_', toUInt64(1)) on node Executing query select dictGetOrDefault('MySQL_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select 20 on node1 Executing query select dictGetOrDefault('RedisSimple_direct_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetDateOrDefault('MySQL_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGet('LocalClickHouse_direct_', 'Int16_', toUInt64(1)) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 [gw9] PASSED test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables Executing query create table test.table_0 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetStringOrDefault('RedisSimple_direct_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGet('MySQL_cache_', 'DateTime_', toUInt64(2)) on node Executing query select dictGetInt16('LocalClickHouse_direct_', 'Int16_', toUInt64(1)) on local_node test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Executing query select dictGet('RedisSimple_direct_String_', 'String_', toUInt64(2)) on node Executing query select dictGetDateTime('MySQL_cache_', 'DateTime_', toUInt64(2)) on node Executing query create table test.table_1 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGetString('RedisSimple_direct_String_', 'String_', toUInt64(2)) on node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query CREATE TABLE mydb.tbl (x UInt8, y String) ENGINE=ReplicatedMergeTree ORDER BY tuple() on node1 Executing query create table test.table_2 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictHas('RedisSimple_direct_String_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query create table test.table_3 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetInt16OrDefault('LocalClickHouse_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetOrDefault('RedisSimple_direct_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query ALTER TABLE mydb.tbl MODIFY COLUMN x String on node1 Connection dropped: socket connection error: None Executing query select dictGetDateTimeOrDefault('MySQL_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query create table test.table_4 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGet('LocalClickHouse_direct_', 'Int32_', toUInt64(1)) on local_node Executing query select dictGetStringOrDefault('RedisSimple_direct_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '29') on node2 Executing query select dictGet('MySQL_cache_', 'String_', toUInt64(2)) on node Executing query create table test.table_5 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetInt32('LocalClickHouse_direct_', 'Int32_', toUInt64(1)) on local_node Executing query system reload dictionary RedisHash_complex_key_hashed_String_ on node Executing query select dictGetString('MySQL_cache_', 'String_', toUInt64(2)) on node Executing query create table test.table_6 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGet('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query create table test.table_7 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetString('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '29') SETTINGS replica_num_in_backup=1 on node1 Executing query select dictHas('RedisHash_complex_key_hashed_String_', tuple(toUInt64(1),toString('world'))) on node Executing query create table test.table_8 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetInt32OrDefault('LocalClickHouse_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetStringOrDefault('MySQL_cache_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query create table test.table_9 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGet('MySQL_cache_', 'Float32_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'Int64_', toUInt64(1)) on local_node Executing query select dictGetStringOrDefault('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query create table test.table_10 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node1 Executing query select dictGetFloat32('MySQL_cache_', 'Float32_', toUInt64(2)) on node Executing query select dictGetInt64('LocalClickHouse_direct_', 'Int64_', toUInt64(1)) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query create table test.table_11 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node2 Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGetString('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query create table test.table_12 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictHas('RedisHash_complex_key_hashed_String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query create table test.table_13 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetFloat32OrDefault('MySQL_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '29') SETTINGS replica_num_in_backup=2 on node2 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGet('MySQL_cache_', 'Float64_', toUInt64(2)) on node Executing query create table test.table_14 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetStringOrDefault('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGetFloat64('MySQL_cache_', 'Float64_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'UUID_', toUInt64(1)) on local_node Executing query create table test.table_15 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node1 Executing query system reload dictionary RedisHash_complex_key_cache_String_ on node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetUUID('LocalClickHouse_direct_', 'UUID_', toUInt64(1)) on local_node Executing query create table test.table_16 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node2 Executing query select dictGet('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('MySQL_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query create table test.table_17 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetString('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat64OrDefault('MySQL_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query create table test.table_18 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictHas('RedisHash_complex_key_cache_String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('MySQL_cache_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query create table test.table_19 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGetUInt64('MySQL_cache_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'Date_', toUInt64(1)) on local_node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node1-1 bash -c ps -C clickhouse] Executing query select dictGetStringOrDefault('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Stdout: PID TTY TIME CMD Stdout: 1577 ? 00:00:09 clickhouse run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw9-node1-1 bash -c pkill clickhouse] Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1577 Executing query select dictGet('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDate('LocalClickHouse_direct_', 'Date_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetString('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictHas('RedisHash_complex_key_cache_String_', tuple(toUInt64(2),toString('qwerty2'))) on node test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetHierarchy('MySQL_cache_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetDateOrDefault('LocalClickHouse_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetHierarchy('MySQL_cache_', toUInt64(2)) on node Stdout:1577 Executing query CREATE TABLE mydb.test (`x` UInt32) ENGINE = ReplicatedMergeTree ORDER BY x on node1 Executing query select dictGetStringOrDefault('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGet('LocalClickHouse_direct_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictIsIn('MySQL_cache_', toUInt64(1), toUInt64(2)) on node Executing query INSERT INTO mydb.test SELECT number AS x FROM numbers(10000000) on node1 Executing query system reload dictionary RedisHash_complex_key_direct_String_ on node Executing query select dictGetDateTime('LocalClickHouse_direct_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictIsIn('MySQL_cache_', toUInt64(2), toUInt64(1)) on node Executing query select dictGet('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] Executing query system reload dictionaries on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1577 Executing query select dictGetString('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('MySQL_direct_', 'UInt8_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictHas('RedisHash_complex_key_direct_String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt8('MySQL_direct_', 'UInt8_', toUInt64(1)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query BACKUP DATABASE mydb TO Disk('backups', '30') on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_direct_', 'String_', toUInt64(1)) on local_node Executing query DROP DATABASE mydb SYNC on node1 Executing query select dictGetStringOrDefault('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGetOrDefault('MySQL_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1577 Executing query select dictGetString('LocalClickHouse_direct_', 'String_', toUInt64(1)) on local_node Executing query RESTORE DATABASE mydb FROM Disk('backups', '30') on node1 Executing query select dictGet('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt8OrDefault('MySQL_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_direct_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetString('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty Executing query select dictGetUInt16('MySQL_direct_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictHas('RedisHash_complex_key_direct_String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetStringOrDefault('LocalClickHouse_direct_', 'String_', toUInt64(9999999), toString('hi')) on local_node Stdout:1577 Executing query select dictGetOrDefault('MySQL_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetStringOrDefault('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGet('LocalClickHouse_direct_', 'Float32_', toUInt64(1)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query system reload dictionaries on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetFloat32('LocalClickHouse_direct_', 'Float32_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_direct_', 'UInt32_', toUInt64(1)) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/.env --project-name roottestdictionariesredis-gw2 --file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml stop --timeout 20] [gw2] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[String_] Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGetUInt32('MySQL_direct_', 'UInt32_', toUInt64(1)) on node run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw9-node1-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/427c91323346a58aa8bb829f4f185b6381c8dda005c71d6ea06ddafdd19ecd6c/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/427c91323346a58aa8bb829f4f185b6381c8dda005c71d6ea06ddafdd19ecd6c/json HTTP/1.1" 200 586 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node test_backup_restore_on_cluster/test.py::test_tables_dependency Executing query CREATE DATABASE mydb ON CLUSTER 'cluster3' on node1 Executing query select dictGetFloat32OrDefault('LocalClickHouse_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query CREATE TABLE mydb.src ON CLUSTER 'cluster' (x Int64, y String) ENGINE=MergeTree ORDER BY tuple() on node1 Executing query select dictGet('LocalClickHouse_direct_', 'Float64_', toUInt64(1)) on local_node Executing query select dictGetUInt32OrDefault('MySQL_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query CREATE DICTIONARY mydb.dict ON CLUSTER 'cluster' (x Int64, y String) PRIMARY KEY x SOURCE(CLICKHOUSE(HOST 'localhost' PORT tcpPort() DB 'mydb' TABLE 'src')) LAYOUT(FLAT()) LIFETIME(0) on node1 run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('MySQL_direct_', 'UInt64_', toUInt64(1)) on node Executing query select dictGetFloat64('LocalClickHouse_direct_', 'Float64_', toUInt64(1)) on local_node Stdout:2389 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw9-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw9-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2389 Executing query select 20 on node1 Executing query CREATE TABLE mydb.dist1 (x Int64) ENGINE=Distributed('cluster', 'mydb', 'src') on node1 Executing query select dictGetUInt64('MySQL_direct_', 'UInt64_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query CREATE TABLE mydb.dist2 (x Int64) ENGINE=Distributed(cluster, 'mydb', 'src') on node3 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query CREATE TABLE mydb.clusterfunc1 AS cluster('cluster', 'mydb.src') on node1 Executing query select 20 on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query CREATE TABLE mydb.clusterfunc2 AS cluster(cluster, mydb.src) on node1 Executing query select dictGetFloat64OrDefault('LocalClickHouse_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query CREATE TABLE mydb.clusterfunc3 AS cluster(cluster, 'mydb', 'src') on node1 Executing query select dictGet('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select dictGet('MySQL_direct_', 'Int8_', toUInt64(1)) on node Executing query CREATE TABLE mydb.clusterfunc4 AS cluster(cluster, dictionary(mydb.dict)) on node1 Stderr: zoo1 Skipped - Image is already being pulled by node1 Stderr: zoo2 Skipped - Image is already being pulled by node1 Stderr: zoo3 Skipped - Image is already being pulled by node1 Stderr: node2 Skipped - Image is already being pulled by node1 Stderr: azurite1 Pulling Stderr: node1 Pulling Stderr: node1 Pulled Stderr: f18232174bc9 Pulling fs layer Stderr: cb2bde55f71f Pulling fs layer Stderr: 9d0e0719fbe0 Pulling fs layer Stderr: 6f063dbd7a5d Pulling fs layer Stderr: f9e3e3d8f042 Pulling fs layer Stderr: a39ef2f62dc8 Pulling fs layer Stderr: 9a21c6b23f0e Pulling fs layer Stderr: efeb7b313b67 Pulling fs layer Stderr: 6fef65209747 Pulling fs layer Stderr: 3d377e512a83 Pulling fs layer Stderr: f9e3e3d8f042 Waiting Stderr: a39ef2f62dc8 Waiting Stderr: 9a21c6b23f0e Waiting Stderr: efeb7b313b67 Waiting Stderr: 6fef65209747 Waiting Stderr: 3d377e512a83 Waiting Stderr: 6f063dbd7a5d Waiting Stderr: 9d0e0719fbe0 Downloading [> ] 15.58kB/1.261MB Stderr: f18232174bc9 Downloading [> ] 48.34kB/3.642MB Stderr: 9d0e0719fbe0 Downloading [==========> ] 276.8kB/1.261MB Stderr: cb2bde55f71f Downloading [> ] 506.1kB/50.34MB Stderr: f18232174bc9 Downloading [======> ] 440.6kB/3.642MB Stderr: 9d0e0719fbe0 Verifying Checksum Stderr: 9d0e0719fbe0 Download complete Stderr: cb2bde55f71f Downloading [==> ] 2.03MB/50.34MB Stderr: f18232174bc9 Downloading [======================> ] 1.669MB/3.642MB Stderr: f18232174bc9 Verifying Checksum Stderr: f18232174bc9 Download complete Stderr: f18232174bc9 Extracting [> ] 65.54kB/3.642MB Stderr: cb2bde55f71f Downloading [=====> ] 5.077MB/50.34MB Stderr: f18232174bc9 Extracting [===================================> ] 2.556MB/3.642MB Stderr: f18232174bc9 Extracting [==================================================>] 3.642MB/3.642MB Stderr: f18232174bc9 Extracting [==================================================>] 3.642MB/3.642MB Stderr: cb2bde55f71f Downloading [===========> ] 11.17MB/50.34MB Stderr: f18232174bc9 Pull complete Stderr: cb2bde55f71f Downloading [====================> ] 20.82MB/50.34MB Stderr: cb2bde55f71f Downloading [=================================> ] 33.52MB/50.34MB Stderr: cb2bde55f71f Downloading [=============================================> ] 46.22MB/50.34MB Stderr: cb2bde55f71f Verifying Checksum Stderr: cb2bde55f71f Download complete Stderr: cb2bde55f71f Extracting [> ] 524.3kB/50.34MB Stderr: cb2bde55f71f Extracting [=====> ] 5.243MB/50.34MB Stderr: cb2bde55f71f Extracting [===========> ] 11.53MB/50.34MB Stderr: cb2bde55f71f Extracting [===================> ] 19.4MB/50.34MB Stderr: cb2bde55f71f Extracting [========================> ] 25.17MB/50.34MB Stderr: cb2bde55f71f Extracting [================================> ] 33.03MB/50.34MB Stderr: cb2bde55f71f Extracting [=========================================> ] 41.42MB/50.34MB Stderr: cb2bde55f71f Extracting [==============================================> ] 46.66MB/50.34MB Stderr: cb2bde55f71f Extracting [==============================================> ] 47.19MB/50.34MB Stderr: cb2bde55f71f Extracting [===============================================> ] 48.23MB/50.34MB Stderr: cb2bde55f71f Extracting [================================================> ] 48.76MB/50.34MB Stderr: cb2bde55f71f Extracting [================================================> ] 49.28MB/50.34MB Stderr: cb2bde55f71f Extracting [=================================================> ] 50.33MB/50.34MB Stderr: cb2bde55f71f Extracting [==================================================>] 50.34MB/50.34MB Stderr: cb2bde55f71f Pull complete Stderr: 9d0e0719fbe0 Extracting [=> ] 32.77kB/1.261MB Stderr: 9d0e0719fbe0 Extracting [==================================================>] 1.261MB/1.261MB Stderr: 9d0e0719fbe0 Extracting [==================================================>] 1.261MB/1.261MB Stderr: 9d0e0719fbe0 Pull complete Stderr: a39ef2f62dc8 Downloading [> ] 3.29kB/209.4kB Stderr: f9e3e3d8f042 Downloading [==================================================>] 116B/116B Stderr: f9e3e3d8f042 Download complete Stderr: 6f063dbd7a5d Downloading [==================================================>] 446B/446B Stderr: 6f063dbd7a5d Verifying Checksum Stderr: 6f063dbd7a5d Download complete Stderr: 6f063dbd7a5d Extracting [==================================================>] 446B/446B Stderr: 6f063dbd7a5d Extracting [==================================================>] 446B/446B Stderr: 6f063dbd7a5d Pull complete Stderr: f9e3e3d8f042 Extracting [==================================================>] 116B/116B Stderr: f9e3e3d8f042 Extracting [==================================================>] 116B/116B Stderr: f9e3e3d8f042 Pull complete Stderr: a39ef2f62dc8 Downloading [==================================================>] 209.4kB/209.4kB Stderr: a39ef2f62dc8 Verifying Checksum Stderr: a39ef2f62dc8 Download complete Stderr: a39ef2f62dc8 Extracting [=======> ] 32.77kB/209.4kB Stderr: a39ef2f62dc8 Extracting [==================================================>] 209.4kB/209.4kB Stderr: a39ef2f62dc8 Extracting [==================================================>] 209.4kB/209.4kB Stderr: a39ef2f62dc8 Pull complete Stderr: efeb7b313b67 Downloading [=> ] 15.58kB/458.8kB Stderr: 9a21c6b23f0e Downloading [> ] 15.58kB/794kB Stderr: efeb7b313b67 Downloading [==============================> ] 276.8kB/458.8kB Stderr: 9a21c6b23f0e Downloading [=================> ] 276.8kB/794kB Stderr: 6fef65209747 Downloading [> ] 375.1kB/36.34MB Stderr: efeb7b313b67 Verifying Checksum Stderr: efeb7b313b67 Download complete Stderr: 9a21c6b23f0e Downloading [==================================================>] 794kB/794kB Stderr: 9a21c6b23f0e Verifying Checksum Stderr: 9a21c6b23f0e Download complete Stderr: 9a21c6b23f0e Extracting [==> ] 32.77kB/794kB Stderr: 6fef65209747 Downloading [==> ] 1.506MB/36.34MB Stderr: 9a21c6b23f0e Extracting [======================> ] 360.4kB/794kB Stderr: 6fef65209747 Downloading [=====> ] 4.143MB/36.34MB Stderr: 9a21c6b23f0e Extracting [===================================> ] 557.1kB/794kB Stderr: 6fef65209747 Downloading [============> ] 9.421MB/36.34MB Stderr: 9a21c6b23f0e Extracting [=============================================> ] 720.9kB/794kB Stderr: 9a21c6b23f0e Extracting [==================================================>] 794kB/794kB Stderr: 6fef65209747 Downloading [========================> ] 17.71MB/36.34MB Stderr: 9a21c6b23f0e Pull complete Stderr: efeb7b313b67 Extracting [===> ] 32.77kB/458.8kB Stderr: 6fef65209747 Downloading [======================================> ] 28.26MB/36.34MB Stderr: efeb7b313b67 Extracting [==================================================>] 458.8kB/458.8kB Stderr: 6fef65209747 Verifying Checksum Stderr: 6fef65209747 Download complete Stderr: efeb7b313b67 Extracting [==================================================>] 458.8kB/458.8kB Stderr: efeb7b313b67 Pull complete Stderr: 6fef65209747 Extracting [> ] 393.2kB/36.34MB Stderr: 6fef65209747 Extracting [=> ] 1.18MB/36.34MB Stderr: 6fef65209747 Extracting [==> ] 1.573MB/36.34MB Stderr: 6fef65209747 Extracting [==> ] 1.966MB/36.34MB Stderr: 6fef65209747 Extracting [===> ] 2.753MB/36.34MB Stderr: 6fef65209747 Extracting [====> ] 3.539MB/36.34MB Stderr: 6fef65209747 Extracting [=====> ] 3.932MB/36.34MB Stderr: 6fef65209747 Extracting [=====> ] 4.325MB/36.34MB Stderr: 6fef65209747 Extracting [=======> ] 5.505MB/36.34MB Stderr: 6fef65209747 Extracting [========> ] 5.898MB/36.34MB Stderr: 6fef65209747 Extracting [========> ] 6.291MB/36.34MB Stderr: 6fef65209747 Extracting [=========> ] 6.685MB/36.34MB Stderr: 6fef65209747 Extracting [=========> ] 7.078MB/36.34MB Stderr: 6fef65209747 Extracting [==========> ] 7.471MB/36.34MB Stderr: 6fef65209747 Extracting [==========> ] 7.864MB/36.34MB Stderr: 6fef65209747 Extracting [===========> ] 8.258MB/36.34MB Stderr: 6fef65209747 Extracting [===========> ] 8.651MB/36.34MB Stderr: 6fef65209747 Extracting [============> ] 9.044MB/36.34MB Stderr: 6fef65209747 Extracting [=============> ] 9.83MB/36.34MB Stderr: 6fef65209747 Extracting [=================> ] 12.98MB/36.34MB Stderr: 6fef65209747 Extracting [==================> ] 13.76MB/36.34MB Stderr: 6fef65209747 Extracting [====================> ] 14.55MB/36.34MB Stderr: 6fef65209747 Extracting [=====================> ] 15.34MB/36.34MB Stderr: 6fef65209747 Extracting [======================> ] 16.12MB/36.34MB Stderr: 6fef65209747 Extracting [=======================> ] 17.3MB/36.34MB Stderr: 6fef65209747 Extracting [==========================> ] 19.27MB/36.34MB Stderr: 6fef65209747 Extracting [==================================> ] 24.77MB/36.34MB Stderr: 6fef65209747 Extracting [========================================> ] 29.49MB/36.34MB Stderr: 6fef65209747 Extracting [============================================> ] 32.64MB/36.34MB Stderr: 6fef65209747 Extracting [==================================================>] 36.34MB/36.34MB Stderr: 6fef65209747 Pull complete Stderr: 3d377e512a83 Downloading [==================================================>] 2.862kB/2.862kB Stderr: 3d377e512a83 Verifying Checksum Stderr: 3d377e512a83 Download complete Stderr: 3d377e512a83 Extracting [==================================================>] 2.862kB/2.862kB Stderr: 3d377e512a83 Extracting [==================================================>] 2.862kB/2.862kB Stderr: 3d377e512a83 Pull complete Stderr: azurite1 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/log', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/config', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/coordination', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/log', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/config', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/coordination', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/log', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/config', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/coordination'] Command:[docker compose --project-name roottestazureblobstoragezerocopyreplication-gw8 --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Executing query select dictGetUInt64('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select 20 on node1 Executing query CREATE TABLE mydb.clusterfunc5 AS clusterAllReplicas(cluster, dictionary(mydb.dict)) on node1 Executing query select dictGetInt8('MySQL_direct_', 'Int8_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query CREATE TABLE mydb.clusterfunc6 AS cluster('cluster', 'mydb.src') on node3 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query CREATE TABLE mydb.clusterfunc7 AS cluster(cluster, mydb.src) on node3 Executing query select 20 on node1 Stderr:time="2025-04-02T02:20:58Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestazureblobstoragezerocopyreplication-gw8_default Creating Stderr: Network roottestazureblobstoragezerocopyreplication-gw8_default Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Started Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Started Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Started Stderr:time="2025-04-02T02:20:59Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:20:59Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.2.4, port:2181, use_ssl:False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select dictGetOrDefault('MySQL_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query CREATE TABLE mydb.clusterfunc8 AS cluster(cluster, 'mydb', 'src') on node3 Executing query select dictGetUInt64OrDefault('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select count() from test.table_4 on node1 Executing query select dictGetInt8OrDefault('MySQL_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query CREATE TABLE mydb.clusterfunc9 AS cluster(cluster, dictionary(mydb.dict)) on node3 Executing query select dictGet('LocalClickHouse_direct_', 'UInt8_', toUInt64(2)) on local_node Executing query select count() from test.table_0 on node1 Executing query CREATE TABLE mydb.clusterfunc10 AS clusterAllReplicas(cluster, dictionary(mydb.dict)) on node3 Executing query select dictGet('MySQL_direct_', 'Int16_', toUInt64(1)) on node Executing query select dictGetUInt8('LocalClickHouse_direct_', 'UInt8_', toUInt64(2)) on local_node Executing query select count() from test.table_1 on node1 Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster3' TO Disk('backups', '31') on node3 Executing query select dictGetInt16('MySQL_direct_', 'Int16_', toUInt64(1)) on node Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select count() from test.table_12 on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select count() from test.table_15 on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query DROP DATABASE mydb on node3 Executing query select count() from test.table_18 on node1 Executing query select dictGetUInt8OrDefault('LocalClickHouse_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster3' FROM Disk('backups', '31') on node3 Executing query select dictGetInt16OrDefault('MySQL_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGet('LocalClickHouse_direct_', 'UInt16_', toUInt64(2)) on local_node Executing query select count() from test.table_14 on node1 Stderr: Container roottestdictionariesredis-gw2-redis1-1 Stopping Stderr: Container roottestdictionariesredis-gw2-node-1 Stopping Stderr: Container roottestdictionariesredis-gw2-redis1-1 Stopped Stderr: Container roottestdictionariesredis-gw2-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/.env --project-name roottestdictionariesredis-gw2 --file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw2/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml down --volumes] Executing query select dictGet('MySQL_direct_', 'Int32_', toUInt64(1)) on node Executing query select count() from test.table_8 on node1 Executing query select dictGetUInt16('LocalClickHouse_direct_', 'UInt16_', toUInt64(2)) on local_node Executing query SYSTEM FLUSH LOGS ON CLUSTER 'cluster3' on node3 Executing query select count() from test.table_17 on node1 Executing query select dictGetInt32('MySQL_direct_', 'Int32_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select count() from test.table_5 on node1 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Stderr: Container roottestdictionariesredis-gw2-node-1 Stopping Stderr: Container roottestdictionariesredis-gw2-redis1-1 Stopping Stderr: Container roottestdictionariesredis-gw2-node-1 Stopped Stderr: Container roottestdictionariesredis-gw2-node-1 Removing Stderr: Container roottestdictionariesredis-gw2-redis1-1 Stopped Stderr: Container roottestdictionariesredis-gw2-redis1-1 Removing Stderr: Container roottestdictionariesredis-gw2-node-1 Removed Stderr: Container roottestdictionariesredis-gw2-redis1-1 Removed Stderr: Network roottestdictionariesredis-gw2_default Removing Stderr: Network roottestdictionariesredis-gw2_default Removed Cleanup called Docker networks for project roottestdictionariesredis-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestdictionariesredis-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestdictionariesredis-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdictionariesredis-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestdictionariesredis-gw2 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:2 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 2 Executing query select count() from test.table_9 on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select count() from test.table_19 on node1 Executing query select dictGetInt32OrDefault('MySQL_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('LocalClickHouse_direct_', 'UInt32_', toUInt64(2)) on local_node run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select count() from test.table_2 on node1 Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.243367 [ 681 ] {ea4bb43e-20e1-440b-98fd-d881852d4a2a} RestorerFromBackup: Table mydb.src has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGet('MySQL_direct_', 'Int64_', toUInt64(1)) on node Executing query select dictGetUInt32('LocalClickHouse_direct_', 'UInt32_', toUInt64(2)) on local_node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.243423 [ 681 ] {ea4bb43e-20e1-440b-98fd-d881852d4a2a} RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.243576 [ 681 ] {ea4bb43e-20e1-440b-98fd-d881852d4a2a} RestorerFromBackup: Table mydb.dist1 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGetInt64('MySQL_direct_', 'Int64_', toUInt64(1)) on node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.243518 [ 681 ] {ea4bb43e-20e1-440b-98fd-d881852d4a2a} RestorerFromBackup: Table mydb.clusterfunc1 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc2 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc2 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select count() from test.table_6 on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.243618 [ 681 ] {ea4bb43e-20e1-440b-98fd-d881852d4a2a} RestorerFromBackup: Table mydb.clusterfunc2 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc3 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc3 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.243467 [ 681 ] {ea4bb43e-20e1-440b-98fd-d881852d4a2a} RestorerFromBackup: Table mydb.clusterfunc3 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc4 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc4 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.243697 [ 681 ] {ea4bb43e-20e1-440b-98fd-d881852d4a2a} RestorerFromBackup: Table mydb.clusterfunc4 has 1 dependencies: mydb.dict (level 2) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc5 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc5 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select count() from test.table_16 on node1 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.243658 [ 681 ] {ea4bb43e-20e1-440b-98fd-d881852d4a2a} RestorerFromBackup: Table mydb.clusterfunc5 has 1 dependencies: mydb.dict (level 2) run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.244488 [ 1580 ] {a95aa81f-887e-4f9c-a32f-93ebe73be27d} RestorerFromBackup: Table mydb.src has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGetUInt32OrDefault('LocalClickHouse_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.244520 [ 1580 ] {a95aa81f-887e-4f9c-a32f-93ebe73be27d} RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist2 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist2 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select count() from test.table_10 on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.242928 [ 682 ] {9d64db3f-8557-4a2d-a415-53d012c47e02} RestorerFromBackup: Table mydb.dist2 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc6 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc6 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.242852 [ 682 ] {9d64db3f-8557-4a2d-a415-53d012c47e02} RestorerFromBackup: Table mydb.clusterfunc6 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc7 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc7 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGet('LocalClickHouse_direct_', 'UInt64_', toUInt64(2)) on local_node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.242876 [ 682 ] {9d64db3f-8557-4a2d-a415-53d012c47e02} RestorerFromBackup: Table mydb.clusterfunc7 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc8 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc8 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGetInt64OrDefault('MySQL_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select count() from test.table_7 on node1 Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.242792 [ 682 ] {9d64db3f-8557-4a2d-a415-53d012c47e02} RestorerFromBackup: Table mydb.clusterfunc8 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc9 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc9 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.242903 [ 682 ] {9d64db3f-8557-4a2d-a415-53d012c47e02} RestorerFromBackup: Table mydb.clusterfunc9 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc10 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc10 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGet('MySQL_direct_', 'UUID_', toUInt64(1)) on node Executing query select dictGetUInt64('LocalClickHouse_direct_', 'UInt64_', toUInt64(2)) on local_node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:01.242826 [ 682 ] {9d64db3f-8557-4a2d-a415-53d012c47e02} RestorerFromBackup: Table mydb.clusterfunc10 has no dependencies (level 0) Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency Executing query select count() from test.table_13 on node1 Executing query select dictGetUUID('MySQL_direct_', 'UUID_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select count() from test.table_3 on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.2.3, port:2181, use_ssl:False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Executing query select count() from test.table_11 on node1 Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.2.2, port:2181, use_ssl:False Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') Trying to create Azurite instance by command docker compose --project-name roottestazureblobstoragezerocopyreplication-gw8 --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d Command:[docker compose --project-name roottestazureblobstoragezerocopyreplication-gw8 --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d] Executing query select dictGetOrDefault('MySQL_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query drop table test.table_0 sync on node1 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUUIDOrDefault('MySQL_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query drop table test.table_1 sync on node1 Executing query select dictGet('LocalClickHouse_direct_', 'Int8_', toUInt64(2)) on local_node Stderr:time="2025-04-02T02:21:05Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Volume "roottestazureblobstoragezerocopyreplication-gw8_data1-1" Creating Stderr: Volume "roottestazureblobstoragezerocopyreplication-gw8_data1-1" Created Stderr:time="2025-04-02T02:21:05Z" level=warning msg="Found orphan containers ([roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 roottestazureblobstoragezerocopyreplication-gw8-zoo2-1]) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up." Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Started Stderr:time="2025-04-02T02:21:05Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:21:05Z" level=debug msg="otel error" error="" Trying to connect to Azurite Executing query drop table test.table_2 sync on node1 Executing query select dictGet('MySQL_direct_', 'Date_', toUInt64(1)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetInt8('LocalClickHouse_direct_', 'Int8_', toUInt64(2)) on local_node Executing query drop table test.table_3 sync on node1 Executing query select dictGetDate('MySQL_direct_', 'Date_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --project-name roottestbackuprestoreoncluster-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/docker-compose.yml stop --timeout 20] Executing query drop table test.table_4 sync on node1 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query drop table test.table_5 sync on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Request URL: 'http://127.0.0.1:30400/devstoreaccount1/?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '2210c110-0f69-11f0-ac90-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Starting new HTTP connection (1): 127.0.0.1:30400 http://127.0.0.1:30400 "GET /devstoreaccount1/?restype=account&comp=properties HTTP/1.1" 200 0 Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '2210c110-0f69-11f0-ac90-0242ac110002' 'x-ms-request-id': '5ea55bae-a0b7-43d8-a13a-0805ccb4ab89' 'x-ms-version': 'REDACTED' 'date': 'Wed, 02 Apr 2025 02:21:06 GMT' 'x-ms-sku-name': 'REDACTED' 'x-ms-account-kind': 'REDACTED' 'x-ms-is-hns-enabled': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' {'client_request_id': '2210c110-0f69-11f0-ac90-0242ac110002', 'request_id': '5ea55bae-a0b7-43d8-a13a-0805ccb4ab89', 'version': '2025-05-05', 'date': datetime.datetime(2025, 4, 2, 2, 21, 6, tzinfo=datetime.timezone.utc), 'sku_name': 'Standard_RAGRS', 'account_kind': 'StorageV2', 'is_hns_enabled': False} Request URL: 'http://127.0.0.1:30400/devstoreaccount1/?comp=REDACTED&prefix=REDACTED&include=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '221e0cda-0f69-11f0-ac90-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30400 "GET /devstoreaccount1/?comp=list&prefix=azurite-container&include= HTTP/1.1" 200 None Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '221e0cda-0f69-11f0-ac90-0242ac110002' 'x-ms-request-id': '3bee01ee-ed3e-42c1-a80b-418025739c5e' 'x-ms-version': 'REDACTED' 'content-type': 'application/xml' 'Date': 'Wed, 02 Apr 2025 02:21:06 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Request URL: 'http://127.0.0.1:30400/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '22200350-0f69-11f0-ac90-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30400 "GET /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 404 None Response status: 404 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-error-code': 'ContainerNotFound' 'x-ms-request-id': '20ebdff8-7a09-4d8f-b9a3-b357a71f5384' 'content-type': 'application/xml' 'Date': 'Wed, 02 Apr 2025 02:21:06 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' azurite container 'azurite-container' doesn't exist, creating it Request URL: 'http://127.0.0.1:30400/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'PUT' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '222132e8-0f69-11f0-ac90-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30400 "PUT /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 201 0 Response status: 201 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'etag': '"0x2409D8C74BF2D20"' 'last-modified': 'Wed, 02 Apr 2025 02:21:06 GMT' 'x-ms-client-request-id': '222132e8-0f69-11f0-ac90-0242ac110002' 'x-ms-request-id': 'c4abadb1-f2ed-4a36-a25c-3be1afea2264' 'x-ms-version': 'REDACTED' 'Date': 'Wed, 02 Apr 2025 02:21:06 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --project-name roottestazureblobstoragezerocopyreplication-gw8 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --project-name roottestazureblobstoragezerocopyreplication-gw8 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml up -d --no-recreate] Executing query drop table test.table_6 sync on node1 Executing query select dictGetInt8OrDefault('LocalClickHouse_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetDateOrDefault('MySQL_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query drop table test.table_7 sync on node1 Executing query select dictGet('LocalClickHouse_direct_', 'Int16_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_direct_', 'DateTime_', toUInt64(1)) on node Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Started Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.2.6... http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None Executing query drop table test.table_8 sync on node1 http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None Executing query select dictGetDateTime('MySQL_direct_', 'DateTime_', toUInt64(1)) on node Executing query select dictGetInt16('LocalClickHouse_direct_', 'Int16_', toUInt64(2)) on local_node http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None Executing query drop table test.table_9 sync on node1 http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None Executing query drop table test.table_10 sync on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None Executing query drop table test.table_11 sync on node1 Executing query select dictGetDateTimeOrDefault('MySQL_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None Executing query select dictGetInt16OrDefault('LocalClickHouse_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query drop table test.table_12 sync on node1 Executing query select dictGet('MySQL_direct_', 'String_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None Executing query drop table test.table_13 sync on node1 http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None Executing query select dictGet('LocalClickHouse_direct_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGetString('MySQL_direct_', 'String_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/acfcde7ed3ce3adc31c17b003f02db9eab8b35efcda6f86468a55c215bc8e931/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.2.7... http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a0e5279c1c425f3a3973e8a0374ad4ebb05bd8b643e70ff58a8e357b52583070/json HTTP/1.1" 200 None Executing query drop table test.table_14 sync on node1 http://localhost:None "GET /v1.46/containers/a0e5279c1c425f3a3973e8a0374ad4ebb05bd8b643e70ff58a8e357b52583070/json HTTP/1.1" 200 None ClickHouse node2 started Cluster started Executing query select dictGetInt32('LocalClickHouse_direct_', 'Int32_', toUInt64(2)) on local_node Executing query CREATE TABLE blob_storage_table ON CLUSTER test_cluster ( id Int64, data String ) ENGINE=ReplicatedMergeTree('/clickhouse/tables/blob_storage_table', '{replica}') ORDER BY id SETTINGS storage_policy='blob_storage_policy',old_parts_lifetime=1 on node1 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query drop table test.table_15 sync on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query drop table test.table_16 sync on node1 Executing query SELECT COUNT(*) FROM blob_storage_table FORMAT Values on node1 Executing query select dictGetStringOrDefault('MySQL_direct_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query drop table test.table_17 sync on node1 Executing query INSERT INTO blob_storage_table VALUES (0,'data'),(1,'data') on node1 Executing query drop table test.table_18 sync on node1 Executing query select dictGet('MySQL_direct_', 'Float32_', toUInt64(1)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query drop table test.table_19 sync on node1 Executing query select dictGetFloat32('MySQL_direct_', 'Float32_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_direct_', 'Int64_', toUInt64(2)) on local_node Executing query SYSTEM SYNC REPLICA blob_storage_table on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/.env --project-name roottestasyncloaddatabases-gw9 --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node2/docker-compose.yml stop --timeout 20] [gw9] PASSED test_async_load_databases/test.py::test_multiple_tables Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node1 Executing query select dictGetInt64('LocalClickHouse_direct_', 'Int64_', toUInt64(2)) on local_node Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node2 Executing query select dictGetOrDefault('MySQL_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Stderr: Container roottestasyncloaddatabases-gw9-node1-1 Stopping Stderr: Container roottestasyncloaddatabases-gw9-node2-1 Stopping Stderr: Container roottestasyncloaddatabases-gw9-node1-1 Stopped Stderr: Container roottestasyncloaddatabases-gw9-node2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/.env --project-name roottestasyncloaddatabases-gw9 --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw9/node2/docker-compose.yml down --volumes] Executing query select dictGetFloat32OrDefault('MySQL_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Request URL: 'http://127.0.0.1:30400/devstoreaccount1/cont?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '25309186-0f69-11f0-ac90-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30400 "GET /devstoreaccount1/cont?restype=container&comp=list HTTP/1.1" 200 None Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'content-type': 'application/xml' 'x-ms-client-request-id': '25309186-0f69-11f0-ac90-0242ac110002' 'x-ms-request-id': '5b415c89-8805-4447-a040-04d3324a3406' 'x-ms-version': 'REDACTED' 'date': 'Wed, 02 Apr 2025 02:21:11 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Executing query INSERT INTO blob_storage_table VALUES (2,'data'),(3,'data') on node2 Executing query select dictGet('MySQL_direct_', 'Float64_', toUInt64(1)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query SYSTEM SYNC REPLICA blob_storage_table on node1 Executing query select dictGetFloat64('MySQL_direct_', 'Float64_', toUInt64(1)) on node Stderr: Container roottestasyncloaddatabases-gw9-node2-1 Stopping Stderr: Container roottestasyncloaddatabases-gw9-node1-1 Stopping Stderr: Container roottestasyncloaddatabases-gw9-node2-1 Stopped Stderr: Container roottestasyncloaddatabases-gw9-node2-1 Removing Stderr: Container roottestasyncloaddatabases-gw9-node1-1 Stopped Stderr: Container roottestasyncloaddatabases-gw9-node1-1 Removing Stderr: Container roottestasyncloaddatabases-gw9-node2-1 Removed Stderr: Container roottestasyncloaddatabases-gw9-node1-1 Removed Stderr: Network roottestasyncloaddatabases-gw9_default Removing Stderr: Network roottestasyncloaddatabases-gw9_default Removed Cleanup called Docker networks for project roottestasyncloaddatabases-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestasyncloaddatabases-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node2 Docker volumes for project roottestasyncloaddatabases-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestasyncloaddatabases-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select dictGet('LocalClickHouse_direct_', 'UUID_', toUInt64(2)) on local_node Unstopped containers: {} No running containers for project: roottestasyncloaddatabases-gw9 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:4 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 4 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node1 Executing query select dictGetUUID('LocalClickHouse_direct_', 'UUID_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Request URL: 'http://127.0.0.1:30400/devstoreaccount1/cont?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '25fe0c38-0f69-11f0-ac90-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30400 "GET /devstoreaccount1/cont?restype=container&comp=list HTTP/1.1" 200 None Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'content-type': 'application/xml' 'x-ms-client-request-id': '25fe0c38-0f69-11f0-ac90-0242ac110002' 'x-ms-request-id': '74ed5455-e5ae-43a7-af4d-2bc620607040' 'x-ms-version': 'REDACTED' 'date': 'Wed, 02 Apr 2025 02:21:13 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Executing query DROP TABLE blob_storage_table ON CLUSTER test_cluster SYNC on node1 Executing query select dictGetFloat64OrDefault('MySQL_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_direct_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetUInt64('MySQL_direct_', 'ParentKeyField', toUInt64(1)) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --project-name roottestazureblobstoragezerocopyreplication-gw8 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml stop --timeout 20] [gw8] PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication Executing query select dictGetUUIDOrDefault('LocalClickHouse_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_direct_', 'Date_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetDate('LocalClickHouse_direct_', 'Date_', toUInt64(2)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_direct_', 'UInt8_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetUInt8('MySQL_direct_', 'UInt8_', toUInt64(2)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetDateTime('LocalClickHouse_direct_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictGetUInt8OrDefault('MySQL_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_direct_', 'UInt16_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetUInt16('MySQL_direct_', 'UInt16_', toUInt64(2)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'String_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetString('LocalClickHouse_direct_', 'String_', toUInt64(2)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGet('MySQL_direct_', 'UInt32_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGetUInt32('MySQL_direct_', 'UInt32_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_direct_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGet('LocalClickHouse_direct_', 'Float32_', toUInt64(2)) on local_node Executing query select dictGetUInt32OrDefault('MySQL_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetFloat32('LocalClickHouse_direct_', 'Float32_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_direct_', 'UInt64_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGetUInt64('MySQL_direct_', 'UInt64_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --project-name roottestbackuprestoreoncluster-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/docker-compose.yml down --volumes] Executing query select dictGetUInt64OrDefault('MySQL_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGet('LocalClickHouse_direct_', 'Float64_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_direct_', 'Int8_', toUInt64(2)) on node Executing query select dictGetFloat64('LocalClickHouse_direct_', 'Float64_', toUInt64(2)) on local_node Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Removed Stderr: Network roottestbackuprestoreoncluster-gw0_default Removing Stderr: Network roottestbackuprestoreoncluster-gw0_default Removed Cleanup called Docker networks for project roottestbackuprestoreoncluster-gw0 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackuprestoreoncluster-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackuprestoreoncluster-gw0 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestoreoncluster-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select dictGetInt8('MySQL_direct_', 'Int8_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Unstopped containers: {} No running containers for project: roottestbackuprestoreoncluster-gw0 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:4 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 4 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetInt8OrDefault('MySQL_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGet('MySQL_direct_', 'Int16_', toUInt64(2)) on node Executing query select dictGetUInt64('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGetInt16('MySQL_direct_', 'Int16_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetInt16OrDefault('MySQL_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetHierarchy('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_direct_', 'Int32_', toUInt64(2)) on node Executing query select dictGetHierarchy('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGetInt32('MySQL_direct_', 'Int32_', toUInt64(2)) on node Executing query select dictIsIn('LocalClickHouse_direct_', toUInt64(1), toUInt64(2)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictIsIn('LocalClickHouse_direct_', toUInt64(2), toUInt64(1)) on local_node Executing query select dictGetInt32OrDefault('MySQL_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] Executing query system reload dictionaries on local_node Executing query select dictGet('MySQL_direct_', 'Int64_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictGetInt64('MySQL_direct_', 'Int64_', toUInt64(2)) on node Executing query select dictGetUInt8('LocalClickHouse_flat_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetInt64OrDefault('MySQL_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGet('MySQL_direct_', 'UUID_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictGetUUID('MySQL_direct_', 'UUID_', toUInt64(2)) on node Executing query select dictGetUInt16('LocalClickHouse_flat_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetUUIDOrDefault('MySQL_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGet('MySQL_direct_', 'Date_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictGetDate('MySQL_direct_', 'Date_', toUInt64(2)) on node Executing query select dictGetUInt32('LocalClickHouse_flat_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGetDateOrDefault('MySQL_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGet('MySQL_direct_', 'DateTime_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictGetDateTime('MySQL_direct_', 'DateTime_', toUInt64(2)) on node Executing query select dictGetUInt64('LocalClickHouse_flat_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --project-name roottestazureblobstoragezerocopyreplication-gw8 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml down --volumes] Executing query select dictGetDateTimeOrDefault('MySQL_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGet('MySQL_direct_', 'String_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int8_', toUInt64(1)) on local_node Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Removed Stderr: Volume roottestazureblobstoragezerocopyreplication-gw8_data1-1 Removing Stderr: Network roottestazureblobstoragezerocopyreplication-gw8_default Removing Stderr: Volume roottestazureblobstoragezerocopyreplication-gw8_data1-1 Removed Stderr: Network roottestazureblobstoragezerocopyreplication-gw8_default Removed Cleanup called Docker networks for project roottestazureblobstoragezerocopyreplication-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestazureblobstoragezerocopyreplication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query select dictGetString('MySQL_direct_', 'String_', toUInt64(2)) on node Docker volumes for project roottestazureblobstoragezerocopyreplication-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestazureblobstoragezerocopyreplication-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestazureblobstoragezerocopyreplication-gw8 Trying to prune unused networks... Executing query select dictGetInt8('LocalClickHouse_flat_', 'Int8_', toUInt64(1)) on local_node Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:2 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 2 test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed Running tests in /ClickHouse/tests/integration/test_config_hide_in_preprocessed/test.py Cluster start called. is_up=False Docker networks for project roottestconfighideinpreprocessed-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfighideinpreprocessed-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfighideinpreprocessed-gw8 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestconfighideinpreprocessed-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfighideinpreprocessed-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Docker volumes for project roottestconfighideinpreprocessed-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfighideinpreprocessed-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Unstopped containers: {} No running containers for project: roottestconfighideinpreprocessed-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:2 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 2 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_hide_in_preprocessed/configs/config.xml'] to /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/node/database Setup logs dir /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/.env --project-name roottestconfighideinpreprocessed-gw8 --file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/node/docker-compose.yml pull] Executing query select dictGetOrDefault('MySQL_direct_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetStringOrDefault('MySQL_direct_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGet('LocalClickHouse_flat_', 'Int16_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_direct_', 'Float32_', toUInt64(2)) on node Executing query select dictGetInt16('LocalClickHouse_flat_', 'Int16_', toUInt64(1)) on local_node Executing query select dictGetFloat32('MySQL_direct_', 'Float32_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetFloat32OrDefault('MySQL_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int32_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_direct_', 'Float64_', toUInt64(2)) on node Executing query select dictGetInt32('LocalClickHouse_flat_', 'Int32_', toUInt64(1)) on local_node Executing query select dictGetFloat64('MySQL_direct_', 'Float64_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetFloat64OrDefault('MySQL_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int64_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_direct_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictGetInt64('LocalClickHouse_flat_', 'Int64_', toUInt64(1)) on local_node Executing query select dictGetUInt64('MySQL_direct_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetHierarchy('MySQL_direct_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UUID_', toUInt64(1)) on local_node Executing query select dictGetUUID('LocalClickHouse_flat_', 'UUID_', toUInt64(1)) on local_node Executing query select dictGetHierarchy('MySQL_direct_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictIsIn('MySQL_direct_', toUInt64(1), toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictIsIn('MySQL_direct_', toUInt64(2), toUInt64(1)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] Executing query system reload dictionaries on node Executing query select dictGet('LocalClickHouse_flat_', 'Date_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt8_', toUInt64(1)) on node Executing query select dictGetDate('LocalClickHouse_flat_', 'Date_', toUInt64(1)) on local_node Executing query select dictGetUInt8('MySQL_flat_', 'UInt8_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetUInt8OrDefault('MySQL_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGet('LocalClickHouse_flat_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetDateTime('LocalClickHouse_flat_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetUInt16('MySQL_flat_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGet('LocalClickHouse_flat_', 'String_', toUInt64(1)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetString('LocalClickHouse_flat_', 'String_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt32_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetUInt32('MySQL_flat_', 'UInt32_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_flat_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/.env --project-name roottestconfighideinpreprocessed-gw8 --file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/.env --project-name roottestconfighideinpreprocessed-gw8 --file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate] Executing query select dictGet('LocalClickHouse_flat_', 'Float32_', toUInt64(1)) on local_node Executing query select dictGetUInt32OrDefault('MySQL_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetFloat32('LocalClickHouse_flat_', 'Float32_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt64_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Stderr: Network roottestconfighideinpreprocessed-gw8_default Creating Stderr: Network roottestconfighideinpreprocessed-gw8_default Created Stderr: Container roottestconfighideinpreprocessed-gw8-node-1 Creating Stderr: Container roottestconfighideinpreprocessed-gw8-node-1 Created Stderr: Container roottestconfighideinpreprocessed-gw8-node-1 Starting Stderr: Container roottestconfighideinpreprocessed-gw8-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestconfighideinpreprocessed-gw8-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestconfighideinpreprocessed-gw8-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestconfighideinpreprocessed-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None Executing query select dictGetUInt64('MySQL_flat_', 'UInt64_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('MySQL_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Float64_', toUInt64(1)) on local_node http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None Executing query select dictGetUInt64OrDefault('MySQL_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None Executing query select dictGetFloat64('LocalClickHouse_flat_', 'Float64_', toUInt64(1)) on local_node http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None Executing query select dictGet('MySQL_flat_', 'Int8_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/22fccb849daf894abfacd20c34cbbf614575d418ef0149fe7cf77148766b62bf/json HTTP/1.1" 200 None ClickHouse node started Executing query select value from system.server_settings where name ='max_thread_pool_free_size' on node Executing query select dictGetInt8('MySQL_flat_', 'Int8_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select value from system.server_settings where name ='max_table_size_to_drop' on node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select value from system.server_settings where name ='max_partition_size_to_drop' on node Executing query select dictGetOrDefault('MySQL_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select collection from system.named_collections on node Executing query select dictGetUInt64('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select dictGetInt8OrDefault('MySQL_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node run container_id:roottestconfighideinpreprocessed-gw8-node-1 detach:False nothrow:False cmd: ['cat', '/var/lib/clickhouse/preprocessed_configs/config.xml'] Command:[docker exec roottestconfighideinpreprocessed-gw8-node-1 cat /var/lib/clickhouse/preprocessed_configs/config.xml] Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: test Stdout: /var/log/clickhouse-server/clickhouse-server.log Stdout: /var/log/clickhouse-server/clickhouse-server.err.log Stdout: Stdout: 1000M Stdout: 10 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: /var/log/clickhouse-server/stderr.log Stdout: /var/log/clickhouse-server/stdout.log Stdout: true Stdout: Stdout: Stdout: Stdout: Stdout: https://{bucket}.s3.amazonaws.com Stdout: Stdout: Stdout: https://storage.googleapis.com/{bucket} Stdout: Stdout: Stdout: https://{bucket}.oss.aliyuncs.com Stdout: Stdout: Stdout: Stdout: Stdout: Stdout:
Stdout: Access-Control-Allow-Origin Stdout: * Stdout:
Stdout:
Stdout: Access-Control-Allow-Headers Stdout: origin, x-requested-with, x-clickhouse-format, x-clickhouse-user, x-clickhouse-key, Authorization Stdout:
Stdout:
Stdout: Access-Control-Allow-Methods Stdout: POST, GET, OPTIONS Stdout:
Stdout:
Stdout: Access-Control-Max-Age Stdout: 86400 Stdout:
Stdout:
Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 8123 Stdout: Stdout: Stdout: 9000 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 9004 Stdout: Stdout: Stdout: 9005 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 9009 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 10 Stdout: Stdout: Stdout: Stdout: Stdout: false Stdout: Stdout: Stdout: /path/to/ssl_cert_file Stdout: /path/to/ssl_key_file Stdout: Stdout: Stdout: false Stdout: Stdout: Stdout: /path/to/ssl_ca_cert_file Stdout: Stdout: Stdout: none Stdout: Stdout: Stdout: 0 Stdout: Stdout: Stdout: -1 Stdout: -1 Stdout: Stdout: Stdout: false Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: none Stdout: true Stdout: true Stdout: sslv2,sslv3 Stdout: true Stdout: Stdout: Stdout: Stdout: RejectCertificateHandler Stdout: Stdout: Stdout: Stdout: Stdout: true Stdout: true Stdout: sslv2,sslv3 Stdout: true Stdout: Stdout: Stdout: Stdout: RejectCertificateHandler Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 0 Stdout: 2 Stdout: fair_round_robin Stdout: Stdout: Stdout: 1000 Stdout: Stdout: Stdout: 0 Stdout: Stdout: Stdout: Stdout: 10000 Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: 0.9 Stdout: Stdout: Stdout: 4194304 Stdout: Stdout: Stdout: 0 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 8589934592 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: / Stdout: Stdout: false Stdout: Stdout: Stdout: /var/lib/clickhouse/ Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: /var/lib/clickhouse/tmp/ Stdout: Stdout: Stdout: 1 Stdout: 1 Stdout: 1 Stdout: Stdout: Stdout: sha256_password Stdout: Stdout: Stdout: 12 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: /var/lib/clickhouse/user_files/ Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: users.xml Stdout: Stdout: Stdout: Stdout: /var/lib/clickhouse/access/ Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: false Stdout: Stdout: Stdout: 2 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: default Stdout: Stdout: Stdout: custom_ Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: default Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: false Stdout: Stdout: ' | sed -e 's|.*>\(.*\)<.*|\1|') Stdout: wget https://github.com/ClickHouse/clickhouse-jdbc-bridge/releases/download/v$PKG_VER/clickhouse-jdbc-bridge_$PKG_VER-1_all.deb Stdout: apt install --no-install-recommends -f ./clickhouse-jdbc-bridge_$PKG_VER-1_all.deb Stdout: clickhouse-jdbc-bridge & Stdout: Stdout: * [CentOS/RHEL] Stdout: export MVN_URL=https://repo1.maven.org/maven2/com/clickhouse/clickhouse-jdbc-bridge/ Stdout: export PKG_VER=$(curl -sL $MVN_URL/maven-metadata.xml | grep '' | sed -e 's|.*>\(.*\)<.*|\1|') Stdout: wget https://github.com/ClickHouse/clickhouse-jdbc-bridge/releases/download/v$PKG_VER/clickhouse-jdbc-bridge-$PKG_VER-1.noarch.rpm Stdout: yum localinstall -y clickhouse-jdbc-bridge-$PKG_VER-1.noarch.rpm Stdout: clickhouse-jdbc-bridge & Stdout: Stdout: Please refer to https://github.com/ClickHouse/clickhouse-jdbc-bridge#usage for more information. Stdout: ]]> Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 3600 Stdout: Stdout: Stdout: Stdout: 3600 Stdout: Stdout: Stdout: 60 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: system Stdout: query_log
Stdout: Stdout: toYYYYMM(event_date) Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 7500 Stdout: Stdout: 1048576 Stdout: Stdout: 8192 Stdout: Stdout: 524288 Stdout: Stdout: false Stdout: Stdout: Stdout: Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: trace_log
Stdout: Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: Stdout: false Stdout: true Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: query_thread_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: query_views_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: part_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: text_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout: trace Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: metric_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: 1000 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: latency_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: 1000 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: error_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: 1000 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: query_metric_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: 1000 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: asynchronous_metric_log
Stdout: 7000 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: engine MergeTree Stdout: partition by toYYYYMM(finish_date) Stdout: order by (finish_date, finish_time_us, trace_id) Stdout: Stdout: system Stdout: opentelemetry_span_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: Stdout: system Stdout: crash_log
Stdout: Stdout: Stdout: 1000 Stdout: 1024 Stdout: 1024 Stdout: 512 Stdout: true Stdout:
Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: system Stdout: processors_profile_log
Stdout: Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout: event_date + INTERVAL 30 DAY DELETE Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: asynchronous_insert_log
Stdout: Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout: event_date Stdout: event_date + INTERVAL 3 DAY Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: backup_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: s3queue_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: blob_storage_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout: event_date + INTERVAL 30 DAY Stdout:
Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: *_dictionary.*ml Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: *_function.*ml Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: /clickhouse/task_queue/ddl Stdout: Stdout: /clickhouse/task_queue/replicas Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: click_cost Stdout: any Stdout: Stdout: 0 Stdout: 3600 Stdout: Stdout: Stdout: 86400 Stdout: 60 Stdout: Stdout: Stdout: Stdout: max Stdout: Stdout: 0 Stdout: 60 Stdout: Stdout: Stdout: 3600 Stdout: 300 Stdout: Stdout: Stdout: 86400 Stdout: 3600 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: /var/lib/clickhouse/format_schemas/ Stdout: Stdout: Stdout: /usr/share/clickhouse/protos/ Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: false Stdout: Stdout: false Stdout: Stdout: Stdout: https://6f33034cfe684dd7a3ab9875e57b1c8d@o388870.ingest.sentry.io/5226277 Stdout: Stdout: false Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: backups Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: node Stdout: Stdout: Stdout: Stdout: Stdout: 1 Stdout: Stdout: Stdout: Stdout: Etc/UTC Stdout: 0.0.0.0 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: users.xml Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 1 Stdout: 1 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 40000000000 Stdout: Stdout:
Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/.env --project-name roottestconfighideinpreprocessed-gw8 --file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/node/docker-compose.yml stop --timeout 20] [gw8] PASSED test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetInt16('MySQL_flat_', 'Int16_', toUInt64(1)) on node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictGetInt16OrDefault('MySQL_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetUInt8('LocalClickHouse_flat_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Int32_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Connection dropped: socket connection error: None Failed connecting to Zookeeper within the connection retry policy. Executing query select dictGetInt32('MySQL_flat_', 'Int32_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Connection dropped: socket connection error: None Executing query select dictGet('LocalClickHouse_flat_', 'UInt16_', toUInt64(2)) on local_node Stderr: Container roottestconfighideinpreprocessed-gw8-node-1 Stopping Stderr: Container roottestconfighideinpreprocessed-gw8-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/.env --project-name roottestconfighideinpreprocessed-gw8 --file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw8/node/docker-compose.yml down --volumes] Executing query select dictGetInt32OrDefault('MySQL_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetUInt16('LocalClickHouse_flat_', 'UInt16_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Int64_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Stderr: Container roottestconfighideinpreprocessed-gw8-node-1 Stopping Stderr: Container roottestconfighideinpreprocessed-gw8-node-1 Stopped Stderr: Container roottestconfighideinpreprocessed-gw8-node-1 Removing Stderr: Container roottestconfighideinpreprocessed-gw8-node-1 Removed Stderr: Network roottestconfighideinpreprocessed-gw8_default Removing Stderr: Network roottestconfighideinpreprocessed-gw8_default Removed Cleanup called Docker networks for project roottestconfighideinpreprocessed-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfighideinpreprocessed-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfighideinpreprocessed-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfighideinpreprocessed-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfighideinpreprocessed-gw8 Trying to prune unused networks... Executing query select dictGetInt64('MySQL_flat_', 'Int64_', toUInt64(1)) on node Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:2 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 2 Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictGetInt64OrDefault('MySQL_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetUInt32('LocalClickHouse_flat_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'UUID_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGetUUID('MySQL_flat_', 'UUID_', toUInt64(1)) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt64_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetUInt64('LocalClickHouse_flat_', 'UInt64_', toUInt64(2)) on local_node Executing query select dictGetUUIDOrDefault('MySQL_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Date_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetDate('MySQL_flat_', 'Date_', toUInt64(1)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int8_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetInt8('LocalClickHouse_flat_', 'Int8_', toUInt64(2)) on local_node Executing query select dictGetDateOrDefault('MySQL_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'DateTime_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetDateTime('MySQL_flat_', 'DateTime_', toUInt64(1)) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int16_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt16('LocalClickHouse_flat_', 'Int16_', toUInt64(2)) on local_node Executing query select dictGetDateTimeOrDefault('MySQL_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'String_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetString('MySQL_flat_', 'String_', toUInt64(1)) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetInt32('LocalClickHouse_flat_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGetStringOrDefault('MySQL_flat_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Float32_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetFloat32('MySQL_flat_', 'Float32_', toUInt64(1)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int64_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetInt64('LocalClickHouse_flat_', 'Int64_', toUInt64(2)) on local_node Executing query select dictGetFloat32OrDefault('MySQL_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Float64_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetFloat64('MySQL_flat_', 'Float64_', toUInt64(1)) on node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UUID_', toUInt64(2)) on local_node Executing query select dictGetUUID('LocalClickHouse_flat_', 'UUID_', toUInt64(2)) on local_node Executing query select dictGetFloat64OrDefault('MySQL_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('MySQL_flat_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetUInt64('MySQL_flat_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetUUIDOrDefault('LocalClickHouse_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Date_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetDate('LocalClickHouse_flat_', 'Date_', toUInt64(2)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt8_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetUInt8('MySQL_flat_', 'UInt8_', toUInt64(2)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_flat_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetDateTime('LocalClickHouse_flat_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictGetUInt8OrDefault('MySQL_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt16_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetUInt16('MySQL_flat_', 'UInt16_', toUInt64(2)) on node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGet('LocalClickHouse_flat_', 'String_', toUInt64(2)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetString('LocalClickHouse_flat_', 'String_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt32_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetUInt32('MySQL_flat_', 'UInt32_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_flat_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Float32_', toUInt64(2)) on local_node Executing query select dictGetUInt32OrDefault('MySQL_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetFloat32('LocalClickHouse_flat_', 'Float32_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt64_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetUInt64('MySQL_flat_', 'UInt64_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Float64_', toUInt64(2)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetFloat64('LocalClickHouse_flat_', 'Float64_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Int8_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetInt8('MySQL_flat_', 'Int8_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGetUInt64('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGetInt8OrDefault('MySQL_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Int16_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetInt16('MySQL_flat_', 'Int16_', toUInt64(2)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetHierarchy('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetHierarchy('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetInt16OrDefault('MySQL_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictIsIn('LocalClickHouse_flat_', toUInt64(1), toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Int32_', toUInt64(2)) on node Executing query select dictIsIn('LocalClickHouse_flat_', toUInt64(2), toUInt64(1)) on local_node Executing query select dictGetInt32('MySQL_flat_', 'Int32_', toUInt64(2)) on node [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] Executing query system reload dictionaries on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetUInt8('LocalClickHouse_hashed_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictGetInt32OrDefault('MySQL_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'Int64_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetInt64('MySQL_flat_', 'Int64_', toUInt64(2)) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetInt64OrDefault('MySQL_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetUInt16('LocalClickHouse_hashed_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'UUID_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetUUID('MySQL_flat_', 'UUID_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictGetUUIDOrDefault('MySQL_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetUInt32('LocalClickHouse_hashed_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'Date_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetDate('MySQL_flat_', 'Date_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGetUInt32OrDefault('LocalClickHouse_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetUInt64('LocalClickHouse_hashed_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictGetDateOrDefault('MySQL_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'DateTime_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetDateTime('MySQL_flat_', 'DateTime_', toUInt64(2)) on node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGetDateTimeOrDefault('MySQL_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt8('LocalClickHouse_hashed_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'String_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetString('MySQL_flat_', 'String_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetInt8OrDefault('LocalClickHouse_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int16_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetStringOrDefault('MySQL_flat_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetInt16('LocalClickHouse_hashed_', 'Int16_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'Float32_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetFloat32('MySQL_flat_', 'Float32_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int32_', toUInt64(1)) on local_node Executing query select dictGetFloat32OrDefault('MySQL_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetInt32('LocalClickHouse_hashed_', 'Int32_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'Float64_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetFloat64('MySQL_flat_', 'Float64_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGet('LocalClickHouse_hashed_', 'Int64_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetInt64('LocalClickHouse_hashed_', 'Int64_', toUInt64(1)) on local_node Executing query select dictGetFloat64OrDefault('MySQL_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetUInt64('MySQL_flat_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UUID_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetUInt64OrDefault('MySQL_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetUUID('LocalClickHouse_hashed_', 'UUID_', toUInt64(1)) on local_node Executing query select dictGetHierarchy('MySQL_flat_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetHierarchy('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetUUIDOrDefault('LocalClickHouse_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictIsIn('MySQL_flat_', toUInt64(1), toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Date_', toUInt64(1)) on local_node Executing query select dictIsIn('MySQL_flat_', toUInt64(2), toUInt64(1)) on node Executing query select dictGetDate('LocalClickHouse_hashed_', 'Date_', toUInt64(1)) on local_node [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] Executing query system reload dictionaries on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt8_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetUInt8('MySQL_hashed_', 'UInt8_', toUInt64(1)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetDateTime('LocalClickHouse_hashed_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictGetUInt8OrDefault('MySQL_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetUInt16('MySQL_hashed_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'String_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetString('LocalClickHouse_hashed_', 'String_', toUInt64(1)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt32_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetUInt32('MySQL_hashed_', 'UInt32_', toUInt64(1)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_hashed_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Float32_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetFloat32('LocalClickHouse_hashed_', 'Float32_', toUInt64(1)) on local_node Executing query select dictGetUInt32OrDefault('MySQL_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt64_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetUInt64('MySQL_hashed_', 'UInt64_', toUInt64(1)) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Float64_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetFloat64('LocalClickHouse_hashed_', 'Float64_', toUInt64(1)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int8_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetInt8('MySQL_hashed_', 'Int8_', toUInt64(1)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select dictGetInt8OrDefault('MySQL_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetUInt64('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int16_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetInt16('MySQL_hashed_', 'Int16_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictGetInt16OrDefault('MySQL_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetUInt8('LocalClickHouse_hashed_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int32_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetInt32('MySQL_hashed_', 'Int32_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt16_', toUInt64(2)) on local_node Executing query select dictGetInt32OrDefault('MySQL_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetUInt16('LocalClickHouse_hashed_', 'UInt16_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int64_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetInt64('MySQL_hashed_', 'Int64_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictGetInt64OrDefault('MySQL_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetUInt32('LocalClickHouse_hashed_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'UUID_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetUUID('MySQL_hashed_', 'UUID_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt64_', toUInt64(2)) on local_node Executing query select dictGetUUIDOrDefault('MySQL_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetUInt64('LocalClickHouse_hashed_', 'UInt64_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'Date_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetDate('MySQL_hashed_', 'Date_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int8_', toUInt64(2)) on local_node Executing query select dictGetDateOrDefault('MySQL_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetInt8('LocalClickHouse_hashed_', 'Int8_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'DateTime_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetDateTime('MySQL_hashed_', 'DateTime_', toUInt64(1)) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int16_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt16('LocalClickHouse_hashed_', 'Int16_', toUInt64(2)) on local_node Executing query select dictGetDateTimeOrDefault('MySQL_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'String_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetString('MySQL_hashed_', 'String_', toUInt64(1)) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetInt32('LocalClickHouse_hashed_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGetStringOrDefault('MySQL_hashed_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'Float32_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetFloat32('MySQL_hashed_', 'Float32_', toUInt64(1)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int64_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetInt64('LocalClickHouse_hashed_', 'Int64_', toUInt64(2)) on local_node Executing query select dictGetFloat32OrDefault('MySQL_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'Float64_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetFloat64('MySQL_hashed_', 'Float64_', toUInt64(1)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UUID_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetUUID('LocalClickHouse_hashed_', 'UUID_', toUInt64(2)) on local_node Executing query select dictGetFloat64OrDefault('MySQL_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGet('MySQL_hashed_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetUInt64('MySQL_hashed_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Date_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetDate('LocalClickHouse_hashed_', 'Date_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt8_', toUInt64(2)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetUInt8('MySQL_hashed_', 'UInt8_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetDateTime('LocalClickHouse_hashed_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetUInt8OrDefault('MySQL_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt16_', toUInt64(2)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetUInt16('MySQL_hashed_', 'UInt16_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'String_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetString('LocalClickHouse_hashed_', 'String_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt32_', toUInt64(2)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_hashed_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetUInt32('MySQL_hashed_', 'UInt32_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Float32_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetFloat32('LocalClickHouse_hashed_', 'Float32_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetUInt32OrDefault('MySQL_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt64_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetUInt64('MySQL_hashed_', 'UInt64_', toUInt64(2)) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Float64_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetFloat64('LocalClickHouse_hashed_', 'Float64_', toUInt64(2)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int8_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetInt8('MySQL_hashed_', 'Int8_', toUInt64(2)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetUInt64('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGetInt8OrDefault('MySQL_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int16_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetInt16('MySQL_hashed_', 'Int16_', toUInt64(2)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetHierarchy('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetHierarchy('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetInt16OrDefault('MySQL_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictIsIn('LocalClickHouse_hashed_', toUInt64(1), toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int32_', toUInt64(2)) on node Executing query select dictIsIn('LocalClickHouse_hashed_', toUInt64(2), toUInt64(1)) on local_node Executing query select dictGetInt32('MySQL_hashed_', 'Int32_', toUInt64(2)) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/.env --project-name roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/local_node/docker-compose.yml stop --timeout 20] [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetInt32OrDefault('MySQL_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('MySQL_hashed_', 'Int64_', toUInt64(2)) on node Executing query select dictGetInt64('MySQL_hashed_', 'Int64_', toUInt64(2)) on node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetInt64OrDefault('MySQL_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGet('MySQL_hashed_', 'UUID_', toUInt64(2)) on node Executing query select dictGetUUID('MySQL_hashed_', 'UUID_', toUInt64(2)) on node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetUUIDOrDefault('MySQL_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGet('MySQL_hashed_', 'Date_', toUInt64(2)) on node Executing query select dictGetDate('MySQL_hashed_', 'Date_', toUInt64(2)) on node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetDateOrDefault('MySQL_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/local_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/local_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/.env --project-name roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw3/local_node/docker-compose.yml down --volumes] Executing query select dictGet('MySQL_hashed_', 'DateTime_', toUInt64(2)) on node Executing query select dictGetDateTime('MySQL_hashed_', 'DateTime_', toUInt64(2)) on node Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1 Stopped Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1 Removing Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-local_node-1 Removed Stderr: Network roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3_default Removing Stderr: Network roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3_default Removed Cleanup called Docker networks for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw3 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Stdout:2 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 2 Executing query select dictGetOrDefault('MySQL_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetDateTimeOrDefault('MySQL_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGet('MySQL_hashed_', 'String_', toUInt64(2)) on node Executing query select dictGetString('MySQL_hashed_', 'String_', toUInt64(2)) on node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetStringOrDefault('MySQL_hashed_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGet('MySQL_hashed_', 'Float32_', toUInt64(2)) on node Executing query select dictGetFloat32('MySQL_hashed_', 'Float32_', toUInt64(2)) on node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetFloat32OrDefault('MySQL_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('MySQL_hashed_', 'Float64_', toUInt64(2)) on node Executing query select dictGetFloat64('MySQL_hashed_', 'Float64_', toUInt64(2)) on node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetFloat64OrDefault('MySQL_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('MySQL_hashed_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictGetUInt64('MySQL_hashed_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetUInt64OrDefault('MySQL_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetHierarchy('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetHierarchy('MySQL_hashed_', toUInt64(2)) on node Executing query select dictIsIn('MySQL_hashed_', toUInt64(1), toUInt64(2)) on node Executing query select dictIsIn('MySQL_hashed_', toUInt64(2), toUInt64(1)) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/.env --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw5 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml stop --timeout 20] [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1 Stopped Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/.env --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw5 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml down --volumes] Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1 Stopped Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1 Removing Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-node-1 Removed Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1 Stopped Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1 Removing Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw5-mysql80-1 Removed Stderr: Network roottestdictionariesalllayoutsseparatesourcesmysql-gw5_default Removing Stderr: Network roottestdictionariesalllayoutsseparatesourcesmysql-gw5_default Removed Cleanup called Docker networks for project roottestdictionariesalllayoutsseparatesourcesmysql-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestdictionariesalllayoutsseparatesourcesmysql-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestdictionariesalllayoutsseparatesourcesmysql-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdictionariesalllayoutsseparatesourcesmysql-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestdictionariesalllayoutsseparatesourcesmysql-gw5 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 ============================== slowest durations =============================== 177.85s setup test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 127.54s call test_async_load_databases/test.py::test_async_load_system_database 50.55s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] 47.56s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] 45.59s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] 45.52s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] 45.38s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] 45.13s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] 44.93s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] 44.61s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] 43.76s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] 41.98s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] 41.28s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] 41.22s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] 41.08s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] 40.41s call test_async_load_databases/test.py::test_dependent_tables 40.31s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] 39.33s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] 34.77s call test_async_load_databases/test.py::test_dict_get_data 27.08s setup test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] 26.85s call test_async_load_databases/test.py::test_multiple_tables 26.40s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 23.89s setup test_cluster_all_replicas/test.py::test_cluster 23.43s call test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] 22.63s call test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] 22.55s call test_dictionaries_redis/test.py::test_redis_dictionaries[String_] 22.51s call test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] 22.46s teardown test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] 22.42s call test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] 22.26s call test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] 22.24s setup test_analyzer_compatibility/test.py::test_two_new_versions 22.22s setup test_config_substitutions/test.py::test_allow_databases 22.11s call test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] 22.07s call test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] 22.03s call test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] 20.79s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] 19.18s teardown test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 18.26s setup test_backup_restore/test.py::test_attach_partition 17.97s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 17.94s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] 17.92s call test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 17.80s teardown test_backup_restore_on_cluster/test.py::test_tables_dependency 17.75s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] 16.99s setup test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] 15.42s teardown test_config_substitutions/test.py::test_include_config 14.81s setup test_async_load_databases/test.py::test_async_load_system_database 14.75s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 14.42s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings 14.30s setup test_composable_protocols/test.py::test_connections 13.31s setup test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed 13.12s setup test_config_corresponding_root/test.py::test_work 13.04s setup test_attach_table_normalizer/test.py::test_attach_substr 12.79s call test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] 12.26s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_user 11.51s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings 11.30s call test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] 10.66s setup test_config_decryption/test.py::test_successful_decryption_xml 10.48s call test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] 10.21s call test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] 10.20s call test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] 9.96s call test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] 9.89s call test_config_substitutions/test.py::test_config_multiple_zk_substitutions 9.41s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] 9.39s call test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] 8.95s teardown test_analyzer_compatibility/test.py::test_two_new_versions 8.67s call test_backup_restore_on_cluster/test.py::test_replicated_database_async 8.42s call test_backup_restore_on_cluster/test.py::test_tables_dependency 8.27s call test_backup_restore/test.py::test_attach_partition 7.97s call test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] 7.75s call test_backup_restore/test.py::test_replace_partition 7.18s call test_backup_restore_on_cluster/test.py::test_required_privileges 6.97s call test_analyzer_compatibility/test.py::test_two_new_versions 6.72s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[String_] 6.61s call test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 6.59s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 6.41s call test_backup_restore/test.py::test_restore 6.38s setup test_aggregation_memory_efficient/test.py::test_remote 6.38s call test_backup_restore_on_cluster/test.py::test_replicated_database 6.32s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 6.24s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 6.13s call test_backup_restore_on_cluster/test.py::test_mutation 5.91s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] 5.80s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 5.68s teardown test_backup_restore/test.py::test_restore 5.56s teardown test_aggregation_memory_efficient/test.py::test_remote 5.28s call test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 5.22s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 5.10s call test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 5.03s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 4.97s call test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 4.81s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 4.80s call test_backup_restore_on_cluster/test.py::test_system_functions 4.75s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 4.75s call test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 4.71s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 4.65s call test_backup_restore_on_cluster/test.py::test_system_users 4.55s call test_backup_restore_on_cluster/test.py::test_replicated_table 4.39s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 4.37s call test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters 4.32s teardown test_config_decryption/test.py::test_successful_decryption_yaml 4.24s call test_backup_restore_on_cluster/test.py::test_projection 4.06s call test_attach_table_normalizer/test.py::test_attach_substr_restart 4.04s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 3.99s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 3.97s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 3.81s call test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 3.79s call test_backup_restore_on_cluster/test.py::test_file_deduplication 3.66s call test_config_substitutions/test.py::test_allow_databases 3.61s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings 3.35s call test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile 3.07s teardown test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed 3.00s call test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 2.86s call test_aggregation_memory_efficient/test.py::test_remote 2.77s call test_backup_restore_on_cluster/test.py::test_empty_replicated_table 2.66s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 2.40s call test_config_substitutions/test.py::test_include_config 2.39s call test_config_substitutions/test.py::test_config 2.32s teardown test_composable_protocols/test.py::test_proxy_1 2.25s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility 2.18s call test_composable_protocols/test.py::test_proxy_1 2.14s teardown test_async_load_databases/test.py::test_multiple_tables 2.13s teardown test_backup_restore_on_cluster/test.py::test_empty_replicated_table 2.03s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 2.03s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 2.03s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 2.01s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 1.99s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_async 1.99s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 1.98s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 1.94s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters 1.93s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 1.93s teardown test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 1.93s teardown test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 1.93s teardown test_backup_restore_on_cluster/test.py::test_mutation 1.89s teardown test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 1.88s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 1.88s teardown test_backup_restore_on_cluster/test.py::test_required_privileges 1.88s teardown test_backup_restore_on_cluster/test.py::test_system_functions 1.88s teardown test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 1.84s teardown test_backup_restore_on_cluster/test.py::test_projection 1.84s teardown test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 1.84s teardown test_backup_restore_on_cluster/test.py::test_replicated_database 1.83s teardown test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 1.83s teardown test_backup_restore_on_cluster/test.py::test_replicated_table 1.83s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 1.82s teardown test_backup_restore_on_cluster/test.py::test_file_deduplication 1.79s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 1.79s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 1.78s teardown test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 1.78s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 1.64s teardown test_backup_restore_on_cluster/test.py::test_system_users 1.61s teardown test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile 1.52s teardown test_attach_table_normalizer/test.py::test_attach_substr_restart 1.41s call test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed 1.36s call test_config_substitutions/test.py::test_config_merge_from_env_overrides 1.22s call test_cluster_all_replicas/test.py::test_cluster 1.05s call test_composable_protocols/test.py::test_connections 1.05s call test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 1.03s call test_config_substitutions/test.py::test_config_from_env_overrides 0.98s call test_attach_table_normalizer/test.py::test_attach_substr 0.80s call test_cluster_all_replicas/test.py::test_global_in 0.53s call test_config_decryption/test.py::test_successful_decryption_yaml 0.53s call test_config_decryption/test.py::test_successful_decryption_xml 0.02s call test_composable_protocols/test.py::test_http_proxy_1 0.01s setup test_backup_restore_on_cluster/test.py::test_projection 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_async_load_databases/test.py::test_async_load_system_database 0.00s setup test_backup_restore_on_cluster/test.py::test_empty_replicated_table 0.00s setup test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 0.00s teardown test_config_corresponding_root/test.py::test_work 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 0.00s setup test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 0.00s setup test_config_decryption/test.py::test_successful_decryption_yaml 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] 0.00s setup test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] 0.00s setup test_backup_restore_on_cluster/test.py::test_system_users 0.00s setup test_backup_restore_on_cluster/test.py::test_system_functions 0.00s setup test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 0.00s setup test_backup_restore_on_cluster/test.py::test_required_privileges 0.00s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[String_] 0.00s setup test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] 0.00s setup test_backup_restore_on_cluster/test.py::test_tables_dependency 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s setup test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 0.00s teardown test_config_decryption/test.py::test_successful_decryption_xml 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] 0.00s setup test_config_substitutions/test.py::test_config_multiple_zk_substitutions 0.00s setup test_async_load_databases/test.py::test_dependent_tables 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] 0.00s teardown test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] 0.00s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 0.00s teardown test_config_substitutions/test.py::test_allow_databases 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] 0.00s setup test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_async 0.00s setup test_backup_restore/test.py::test_replace_partition 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] 0.00s setup test_attach_table_normalizer/test.py::test_attach_substr_restart 0.00s teardown test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] 0.00s teardown test_cluster_all_replicas/test.py::test_cluster 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table 0.00s setup test_composable_protocols/test.py::test_proxy_1 0.00s teardown test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] 0.00s teardown test_attach_table_normalizer/test.py::test_attach_substr 0.00s setup test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_file_deduplication 0.00s setup test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] 0.00s setup test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] 0.00s setup test_backup_restore_on_cluster/test.py::test_mutation 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] 0.00s setup test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] 0.00s teardown test_cluster_all_replicas/test.py::test_global_in 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] 0.00s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_user 0.00s teardown test_composable_protocols/test.py::test_connections 0.00s call test_config_corresponding_root/test.py::test_work 0.00s setup test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] 0.00s setup test_backup_restore/test.py::test_restore 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility 0.00s setup test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 0.00s teardown test_config_substitutions/test.py::test_config_merge_from_env_overrides 0.00s teardown test_backup_restore/test.py::test_attach_partition 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] 0.00s teardown test_async_load_databases/test.py::test_dependent_tables 0.00s teardown test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] 0.00s teardown test_async_load_databases/test.py::test_dict_get_data 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 0.00s setup test_config_substitutions/test.py::test_config_from_env_overrides 0.00s setup test_async_load_databases/test.py::test_multiple_tables 0.00s setup test_config_substitutions/test.py::test_include_config 0.00s teardown test_backup_restore/test.py::test_replace_partition 0.00s teardown test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] 0.00s teardown test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] 0.00s setup test_config_substitutions/test.py::test_config_merge_from_env_overrides 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] 0.00s setup test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile 0.00s setup test_async_load_databases/test.py::test_dict_get_data 0.00s teardown test_config_substitutions/test.py::test_config_multiple_zk_substitutions 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s setup test_composable_protocols/test.py::test_http_proxy_1 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_user 0.00s setup test_config_substitutions/test.py::test_config 0.00s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility 0.00s setup test_cluster_all_replicas/test.py::test_global_in 0.00s teardown test_config_substitutions/test.py::test_config 0.00s teardown test_composable_protocols/test.py::test_http_proxy_1 0.00s teardown test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings 0.00s teardown test_config_substitutions/test.py::test_config_from_env_overrides 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] =========================== short test summary info ============================ PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] PASSED test_cluster_all_replicas/test.py::test_cluster PASSED test_config_substitutions/test.py::test_allow_databases PASSED test_backup_restore/test.py::test_attach_partition PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] PASSED test_config_substitutions/test.py::test_config PASSED test_config_substitutions/test.py::test_config_from_env_overrides PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings PASSED test_config_substitutions/test.py::test_config_merge_from_env_overrides PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility PASSED test_backup_restore/test.py::test_replace_partition PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] PASSED test_cluster_all_replicas/test.py::test_global_in PASSED test_config_substitutions/test.py::test_config_multiple_zk_substitutions PASSED test_backup_restore/test.py::test_restore PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica PASSED test_config_substitutions/test.py::test_include_config PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_user PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes PASSED test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication PASSED test_config_decryption/test.py::test_successful_decryption_xml PASSED test_config_decryption/test.py::test_successful_decryption_yaml PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host PASSED test_composable_protocols/test.py::test_connections PASSED test_composable_protocols/test.py::test_http_proxy_1 PASSED test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] PASSED test_composable_protocols/test.py::test_proxy_1 PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] PASSED test_attach_table_normalizer/test.py::test_attach_substr PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] PASSED test_backup_restore_on_cluster/test.py::test_mutation PASSED test_attach_table_normalizer/test.py::test_attach_substr_restart PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] PASSED test_backup_restore_on_cluster/test.py::test_projection PASSED test_backup_restore_on_cluster/test.py::test_replicated_database PASSED test_config_corresponding_root/test.py::test_work PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] PASSED test_aggregation_memory_efficient/test.py::test_remote PASSED test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table PASSED test_analyzer_compatibility/test.py::test_two_new_versions PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster PASSED test_async_load_databases/test.py::test_async_load_system_database PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] PASSED test_async_load_databases/test.py::test_dependent_tables PASSED test_backup_restore_on_cluster/test.py::test_required_privileges PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] PASSED test_backup_restore_on_cluster/test.py::test_system_functions PASSED test_backup_restore_on_cluster/test.py::test_system_users PASSED test_async_load_databases/test.py::test_dict_get_data PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[String_] PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency PASSED test_async_load_databases/test.py::test_multiple_tables PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] PASSED test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] ======================= 100 passed in 376.06s (0:06:16) ========================